Iron deficiency and anaemia, which affect mainly young children and women of childbearing age, are longstanding public health issues in the USA. According to the Third National Health and Nutrition Examination Survey (NHANES III), iron deficiency anaemia (IDA) per se affected 3% of children aged 3–5 years and 2–5% of adolescent girls and womenReference Looker, Dallman, Carroll, Gunter and Johnson1.
Fortification is generally considered a cost-effective strategy for preventing IDA2–Reference Yip and Ramakrishnan4. Many micro-level studies and field trials have shown that fortified food products can improve the iron status of individualsReference Torrejon, Castillo-Duran, Hertrampf and Ruz5–Reference Sari, Bloem, de Pee, Schultink and Sastroamidjojo11. However, results from these studies cannot be directly extrapolated to the national level because the subjects, dietary patterns and other conditions in trials may not reflect conditions in the population at large. For instance, a decline in anaemia has been well-documented among infant and child participants in the US Special Supplemental Nutrition Program for Women, Infants, and Children (WIC) from the late 1970s to the 1990s, coinciding with improvements in the quality and quantity of iron in commonly consumed fortified infant foodsReference Yip and Ramakrishnan4, Reference Sherry, Mei and Yip12–Reference Vazquez-Seoane, Windom and Pearson15. However, this finding has not been confirmed in representative national surveysReference Looker, Dallman, Carroll, Gunter and Johnson1, 16. Similarly, macro-level studies based on nationally representative samples are difficult to interpret because fortification or defortification may coincide with other significant changes in food supply, dietary patterns, supplement use and medical care during the same periodReference Milman, Byg, Ovesen, Kirchhoff and Jurgensen17, Reference Milman, Byg, Ovesen, Kirchhoff and Jurgensen18. The present study employs a method that builds upon the strengths of micro- and macro-level approaches to examine the impact of national fortification policy in the USA.
The need for national-level estimates of the impact of fortification policy arises for two reasons. First, there is great interest in fortification as a possible strategy for reducing iron deficiency in developing countriesReference Yip and Ramakrishnan4. Second, in developed countries the discovery of genetic subpopulations at risk for iron overload, combined with lower overall levels of iron deficiency than in earlier periods, has raised concerns about the continuation of universal fortification policiesReference Osler, Milman and Heitmann19, Reference Olsson, Vaisanen, Konar and Bruce20.
This paper presents the results of a novel method for estimating the national-level effects of iron defortification in the USA. The method, based on Monte Carlo simulation, exploits the strengths of micro- and macro-level approaches by integrating information from varied sources, while also retaining the capacity to separate the effects of iron defortification from other features of the food supply and dietary patterns.Footnote * The utility of simulation and other systems-based approaches in public health has been noted recentlyReference McLeroy21.
Research design and methods
Model overview
We developed a Monte Carlo simulation model using @Risk 4.5 for Excel (Palisades Inc.). Monte Carlo simulation is a computer simulation technique that uses sampling from a random number sequence to simulate characteristics or events with probability distributions. A Monte Carlo model of a population is composed of hundreds or thousands of individuals whose aggregated characteristics follow certain distributions. So the simulated model preserves inter-individual variations within the population and allows population-level outcomes to be expressed in probabilistic terms.
The model used for this study generated a representative cohort of the US population composed of 15 000 individuals using Monte Carlo sampling. The dynamic changes in total body iron (TBI) of each individual during a 15-year follow-up period were simulated. We tested the impact of iron fortification on iron status and the prevalence of IDA in the cohort.
TBI in the model was composed of three components: iron in haemoglobin (Hb), non-storage iron in tissues and storage iron (Fig. 1). The small amount of transport iron in plasma and extracellular fluid was neglected in the model. During an individual's life, TBI is a changing variable. TBI at a certain time point (t i+1) in the model is a function of previous iron status (t i), iron loss and iron absorption during the time period between t i and t i+1. Data sources used to define TBI at t 0, and parameters used to simulate iron absorption and iron loss during various time periods, are listed in Table 1Reference Kuczmarski, Ogden, Guo, Grummer-Strawn, Flegal and Mei22–Reference Ventura, Mosher, Curtin, Abma and Henshaw32. Model structure is described in detail below.
Hb – haemoglobin; SF – serum ferritin; NHANES III – Third National Health and Nutrition Examination Survey; CSFII – Continuing Survey of Food Intakes by Individuals.
Model structure
To define TBI at baseline (t 0), iron in Hb was calculated based on Hb level and total blood volume estimated from weight and height, and non-storage iron in tissues estimated from body weight24, Reference Linderkamp, Versmold, Riegel and Betke33–Reference Nadler and Bloch35. Storage iron in tissues was calculated from serum ferritin (SF) level according to the well-documented linear relationship between SF and storage ironReference Hallberg and Rossander-Hulten36–Reference Hallberg, Hulthen and Garby38.
To estimate each individual's TBI at t 0, his/her age, gender, height, weight, Hb and SF levels must be defined. NHANES III data were used to generate the empirical distributions of age and gender ratios for the US population. The population was divided into the following age groups when necessary: < 3, 3–6, 6–11, 12–16, 17–20, 21–29, 30–49, 50–59, 60–69 and ≥ 70 years. For each age and gender group, the distributions of height, weight, Hb and SF were calculated using NHANES III data. Correlation coefficient matrices from the empirical NHANES III data were used to more accurately model the relationships among input variables that were not independent of one another. Monte Carlo sampling was used to generate specific values for these variables for each individual in the model according to these distributions. This procedure generated the starting TBI level for each individual in the cohort.
Two time units were used to facilitate the simulation of changes in iron status once individuals left baseline (t 0) and were projected into the future. The first unit was the month. Events simulated at monthly intervals include iron absorption from food and supplements, daily iron loss and menstruation. Within one simulation month, each individual was assumed to follow the same dietary intake and supplement intake level if they took any iron supplement. Data from the Continuing Survey of Food Intakes by Individuals (CSFII) 1994–96 were used to estimate meat, iron, calcium and ascorbic acid intakes for each gender and age group. Data were normalised using Cox–Box transformations, then analysis of variance was used to estimate average dietary intake levels, intra- and inter-individual variations39. Using Monte Carlo sampling, specific values were assigned to each individual according to these distributions. An individual's mean intake level of a nutrient for the month was assigned based on the average intake and inter-individual variance for his/her subgroup, then daily intake values were assigned based on each person's mean intake level and intra-individual variance. Linear equations were used to estimate the percentage of dietary iron from fortification based on iron intake from previous workReference Berner, Clydesdale and Douglass23. The same Monte Carlo technique was used to determine whether an individual in the model takes supplements, the average daily dose taken, and the length of time that this individual will take iron supplements based on the frequency, dosage and length of iron supplement usage reported in NHANES III.
Iron absorption was determined by meal composition and the individual's iron status. Monsen and Balintfy's model and the regression equation of Hallberg et al. were used to calculate bioavailable iron based on iron intake and the amount of iron-absorption enhancers (ascorbic acid and meat) and iron-absorption inhibitors (calcium) contained in the dietReference Monsen and Balintfy28, Reference Hallberg, Brune, Erlandsson, Sandberg and Rossander-Hulten29. Data on iron absorption among normal individuals reported by Lynch et al. were analysed to generate the linear regression equations on log iron absorption rates and SF levelsReference Lynch, Skikne and Cook30. The equations were used to further adjust the haem and non-haem iron absorption according to each individual's iron store as indicated by SF level. Iron from supplements was assumed to enter the non-haem iron pool in the small intestine, thus its bioavailability was assumed to be the same as non-haem iron from food.
Daily iron loss was assumed to be 14 μg per kg24. Average daily loss was calculated according to body weight and iron loss per unit body weight. For adolescent girls older than 11 years and for women of reproductive age in the cohort, the amount of menstrual blood loss per menstrual cycle was assigned based on an empirical distributionReference Hallberg, Hogdahl, Nilsson and Rybo27. The amount of iron lost each month during menstruation was calculated based on the amount of blood loss and Hb level. At the end of each month, storage iron was updated, taking into account iron absorption and iron loss during the month. Consequently, the SF level was updated corresponding to the new amount of storage iron.
Year was the second time unit in the simulation model. Events simulated on a yearly interval include growth, blood donations and pregnancy. For children and adolescents, the expected growth was estimated using the 2000 growth curves of the Centers for Disease Control and PreventionReference Kuczmarski, Ogden, Guo, Grummer-Strawn, Flegal and Mei22. Because of increased height and weight, iron in body tissues expanded and Hb iron also increased because of increased blood volume and Hb levels.
Pregnancy was simulated as an event happening within a given simulation year. Iron loss during pregnancy was based on experts' judgementReference Bothwell, Cook and Finch25, Reference Bothwell26, Reference Hallberg and Berger40. The most commonly used distribution to quantify experts' opinions, triangle distribution, was applied. The small loss of iron from breast milk was neglected. The probabilities of giving birth were estimated from national vital statistics and varied according to mother's ageReference Ventura, Mosher, Curtin, Abma and Henshaw32. Iron supplement usage during pregnancy, the length of taking supplements and the amount of iron taken were simulated based on data from NHANES III. NHANES III data were also used to calculate the probability of being a blood donor for each age and gender group and the frequency of blood donation each year. For blood donors in the simulation model, iron loss through blood donation was calculated using frequency of blood donation per year, units of blood donated each time and Hb level.
At the end of each simulation year, TBI, Hb and SF levels were updated as the result of these events. Distributions of TBI were recalculated following the rule that iron in the body is used to build tissue and maintain normal Hb level first before extra iron is storedReference Brock, Halliday, Pippard and Powell41. Iron in the body was first allocated to body tissues, then to maintain a normal Hb level. Only after meeting these two needs was the remaining iron apportioned to storage. If the amount of TBI was insufficient to maintain a normal Hb level, the highest Hb level possible was calculated.
Mortality rates were obtained from national vital statistics reports. The mortality rates for different gender, age and ethnic groups in 1994 were used in the model31. Based on these rates, mortality in various subgroups was randomly assigned.
Verification and validation
As verification, the computer program was debugged to eliminate logical errors and ensure that the parameters entered into the model were correct. Intermediate simulation outputs, such as simulated one-day nutrient intakes and simulated iron absorption rates, were checked to ensure that the intermediate outputs of the simulation model conformed to general expectations.
Baseline iron status (SF, Hb and TBI at t 0) of the cohort was directly generated from national survey, so it represents the iron status of the US population. However, once individuals in the simulation cohort left baseline, TBI at future times was the result of previous TBI, the algorithm and the parameters used in the model to calculate iron absorption and iron requirements during the interval. Thus, since iron nutrition at the population level in the USA is not subject to dramatic changes from year to year, an important element of validation is whether the model can reproduce Hb and SF distributions at the end of simulation years. Hb and SF levels for each individual at the end of every simulation year were recorded in a data file. The distributions of Hb and SF for different gender and age groups in the simulation cohort were calculated and compared with historical data. Since longitudinal data on iron status is very limited, and estimates for the dietary and other input variables reflect conditions in the mid-1990s, we chose to use NHANES III data for validation.
Results
Validation
In the simulation model, the median TBI after 15 simulation years was 3.4 g for adult men and 2.2 g for adult women. This amount is consistent with estimates from experimentsReference Bothwell, Cook and Finch25. Hb and SF were the two other iron status indicators in the simulation model. Figure 2 compares the distributions of Hb and SF from the simulation model with empirical data from NHANES III.
Compared with NHANES III data, the difference between simulated and observed Hb distributions was small. At the 5th and 50th percentiles, the differences between the simulated SF and observed SF were between 0 and 40 μg l− 1 across age and gender groups. The major difference between simulated SF and empirical data was at the 95th percentile. The largest difference was 210 μg l− 1 among females and 343 μg l− 1 among males. In general, variations in SF in the simulation cohort were less than in the empirical distributions. This may be caused by two factors. First, the simulation model was a simplified version of reality. Some factors that can influence SF levels in the real world were not simulated in the model (e.g. medical treatment, chronic disease and pathological losses of iron). Second, the mismatch between simulation and NHANES III data was most obvious among those with high iron storage, i.e. high SF levels. This may be related to the existence of genetic subpopulations in the USA that were not simulated in this paper, as discussed later.
Iron defortification
With the iron defortification model, there was no significant decrease of Hb among women at the 50th and 95th percentiles (Fig. 3). However, at the 5th percentile, Hb level decreased by 1.5 to 4 g l− 1 among those aged 1–6 years and by 7 to 10 g l− 1 among those aged 10–60 years. This suggests that the majority of women could maintain normal Hb levels without iron fortification, but a small percentage at the low end of the distribution would experience a decrease in Hb levels when their iron intake decreases due to defortification. Among males, there was no detectable change in Hb distributions after iron fortification.
Figure 3 shows a systematic decrease in SF levels among both males and females with iron defortification. Among children, the decrease in SF ranged from 4 to 13 μg l− 1; among adolescent boys and adult men, from 10 to 20 μg l− 1.
The prevalence of IDA was estimated based on the simulated Hb and SF values. With iron fortification, the prevalence of IDA in the simulation model was similar to the prevalence of IDA observed in NHANES III data (Table 2). A difference of approximately 2% in the prevalence was found between the simulation model and NHANES III among children less than 3 years old, in females between 12 and 49 years of age and among adults older than 50 years. Without iron fortification, the increase in IDA was most significant among children and women of reproductive age. Among children less than 11 years old, the prevalence of IDA increased by about 2 to 4.5%. For girls between 12 and 15 years of age, the prevalence increased by 3.8%; similarly, for women between 20 and 49 years, the prevalence increased by 3.7%. The increase was most dramatic (more than 5%) among 16- to 19-year-old girls. Further analysis (not shown) revealed that 45–89% of the excess cases of IDA in various age groups are in the mild category (Hb>100 g l− 1), 10–39% are in the moderate category (Hb = 70–100 g l− 1) and less than 17% are severe (Hb < 70 g l− 1)Reference Tao42. Adolescent boys, adult men and women after menopause were at lowest risk for IDA, with prevalence of less than 1% after iron defortification.
IDA – iron-deficiency anaemia; NHANES III – Third National Health and Nutrition Examination Survey; CI – confidence interval.
* Data source: Reference Looker, Dallman, Carroll, Gunter and Johnsonref. 1.
Discussion
The primary findings from this study are that, under the dietary patterns, nutrient intake levels and other conditions in the mid-1990s in the USA, iron defortification would: (1) lead to a modest decrease in iron stores in all age/sex groups, as seen in SF levels – this effect is seen at all centiles of the distribution examined; (2) decrease Hb levels (and increase IDA levels) among the most vulnerable groups (young children and adolescent and reproductive-aged women) – this effect is seen only at the lower end of the distribution in these groups, where the need is greatest; and (3) have smaller but detectable effects in increasing IDA among prepubescent children and adolescent boys, with no effect on the Hb levels of adult men.
As with all simulations, the validity of these results depends upon the accuracy of the input variables, the parameters defining key relationships among the variables and the structure of the biological model. In the present case, the strengths of the study include the use of nationally representative data to estimate the distributions for many of the input variables and the existence of a substantial literature concerning the biology of iron nutrition from which key model parameters were derived. In addition, several aspects of the empirical results provide some confidence in the model. First, the model is able to reproduce the observed national distributions of SF and Hb (Fig. 2). Second, the one notable discrepancy between simulated and observed distributions (i.e. the 95th percentile for SF in Fig. 2) has a plausible explanation based on current knowledge. Specifically, the SF distribution observed in NHANES III data reflects the presence of genetic subpopulations at risk for iron overload, but these subpopulations were not included in the simulation model presented here. Although less than 0.5% of the US population is homozygous for the C282Y mutation (the variant most strongly associated with haemochromatosis), approximately 9% of the population is heterozygous for this mutationReference Hanson, Imperatore and Burke43 and have been shown to have elevated SF and transferrin saturation levelsReference Mura, Le Gac, Raguenes, Mercier, Le Guen and Ferec44, Reference Cogswell, Gallagher, Steinberg, Caudill, Looker and Bowman45.
It is noteworthy that the present simulation model produces plausible and coherent results despite some limitations in input variables and parameter estimates. For instance: (1) the nutrient intake distributions were adjusted for intra-individual variation using standard methods39 but it was not possible to model how individuals' relative position in the distribution may change during their lifetime; (2) fractional iron absorption was adjusted for some known influences (ascorbic acid, meat, calcium and haem iron) but not others (e.g. phytate, tannins); and (3) fractional absorption rates (and inter-individual variation in these rates) were estimated from a relatively small number of studies with a limited range of variation in age, gender, meal conditions and iron status. The ability of the model to produce plausible results despite these limitations may relate to the strong homeostatic properties of iron status across a wide range of intakes, the focus on population- rather than individual-level changes and the inclusion of the major determinants of iron status in the model (with the exception of genetic variants).
Our simulation results differ from an observational study in Denmark, which reported iron nutrition among Danish adults above the age of 40 years before and after Denmark ended its mandatory iron fortification policy in 1987. That study is important because few researchers have used national representative samples to address this issue and few countries have undertaken defortification. Unlike our simulation results, the Danish study found no increase in iron deficiency after defortification. Among menopausal women and adult men, SF levels even showed some increaseReference Milman, Byg, Ovesen, Kirchhoff and Jurgensen17, Reference Milman, Byg, Ovesen, Kirchhoff and Jurgensen18, Reference Olsson, Vaisanen, Konar and Bruce20, Reference Milman, Pedersen, Ovesen and Schroll46, Reference Milman, Byg and Ovesen47. However, the impact of confounders could not be excluded in these observational studies. As the authors noted, the decrease in iron intake after defortification coincided with less consumption of dairy products, eggs and tea, and greater consumption of meat and poultryReference Milman, Byg, Ovesen, Kirchhoff and Jurgensen17, Reference Milman, Byg, Ovesen, Kirchhoff and Jurgensen18. These concomitant dietary changes may have compensated for the decrease in iron intake due to defortification or, alternatively, population iron status may not be as strongly influenced by fortification as is generally assumed. These confounding factors are a major reason why observational studies provide a limited basis for evaluating the effectiveness of national fortification programmes and illustrate the utility of microsimulation methods as a means for separating the effects of fortification or defortification from other changes at the national level.
In addition to these methodological differences, the US and Danish populations differ in key aspects of their diet and food supply. The US population has substantially higher dietary iron intake and iron fortification levels, with a much lower percentage of women taking iron supplementsReference Milman, Byg, Ovesen, Kirchhoff and Jurgensen17, Reference Cogswell, Kettel-Khan and Ramakrishnan48. Thus food appears to be a more important source of iron in the USA than in Denmark, and iron fortification represents a higher fraction of total iron intake, such that fortification may be a more important determinant of iron nutrition in the USA than in Denmark.
Sweden is another country that withdrew iron fortification of flour in 1995. The preliminary result from Sweden is more consistent with our simulation results. One study in Swedish schools showed that iron deficiency increased from 39.6% in 1994 to 48.9% (P < 0.001) in year 2000 among adolescent girls and iron defortification was suggested to be the main cause for the increase iron deficiencyReference Hulthén and Hallberg49. Further results from the Swedish study are needed to understand the similarities and differences between the Swedish and Danish experiences and the simulated model of the US population.
Conclusion
Estimating the effects of fortification policy is problematic because neither micro- nor macro-level studies, by themselves, provide firm answers. This paper improves upon previous studies by using a method that integrates information from micro- and macro-level studies and retains the capacity to separate the effects of iron fortification from other features of the food supply and dietary patterns. The study suggests that defortification in the USA would increase IDA by 2 to 5 percentage points among pre-school and prepubescent children, adolescent girls and women of reproductive age. However, in the majority of women, children and men, defortification would not change Hb levels and would decrease iron stores by an amount unlikely to have public health implications.
It is important to note that the present simulation does not examine the potential influence of changes in health care, the WIC programme, targeted fortification or other factors in preventing or ameliorating the rise in IDA among vulnerable groups, nor the cost-effectiveness of such approaches compared with fortification. The policy implications of the present findings depend upon such considerations and could be informed through further simulations that examine these issues. In addition, the policy implications depend on the extent to which defortification might slow the progression of iron overload and its sequelae, as discussed in a separate paperReference Tao42.
Acknowledgements
Sources of funding: M.T. was funded by scholarship from the Division of Nutritional Sciences, Cornell University. This work was partly funded by an unrestricted education grant to Cornell University by the Merck Company Foundation, the philanthropic arm of Merck & Co. Inc.
Conflict of interest declaration: The authors have no conflicts of interest.
Authorship responsibilities: M.T. developed the simulation model and prepared the manuscript. D.L.P. initiated the overall project of which this paper is a part, worked closely with M.T. in developing the conceptual underpinnings of the model and interpreting model results, and participated in drafting the paper. D.D.M. worked with M.T. in developing the model and interpreting model results, and participated in preparing the paper.