Hostname: page-component-78c5997874-lj6df Total loading time: 0 Render date: 2024-11-10T16:19:47.409Z Has data issue: false hasContentIssue false

Iron deficiency anaemia: experiences and challenges

Published online by Cambridge University Press:  10 July 2018

Isabella Stelle
Affiliation:
Institute of Liver and Digestive Health, Division of Medicine, University College London, London WC1E 6BT, UK
Anastasia Z. Kalea
Affiliation:
Institute of Liver and Digestive Health, Division of Medicine, University College London, London WC1E 6BT, UK Institute of Cardiovascular Science, University College London, London WC1E 6DD, UK School of Human Sciences, London Metropolitan University, London N7 8DB, UK
Dora I. A. Pereira*
Affiliation:
Department of Pathology, University of Cambridge, Cambridge CB2 1QP, UK MRC Unit The Gambia at the London School of Hygiene & Tropical Medicine, Fajara, Banjul, The Gambia
*
*Corresponding author: Dr D. Pereira, email diap2@cam.ac.uk
Rights & Permissions [Opens in a new window]

Abstract

Iron deficiency remains the largest nutritional deficiency worldwide and the main cause of anaemia. Severe iron deficiency leads to anaemia known as iron deficiency anaemia (IDA), which affects a total of 1·24 billion people, the majority of whom are children and women from resource-poor countries. In sub-Saharan Africa, iron deficiency is frequently exacerbated by concomitant parasitic and bacterial infections and contributes to over 120 000 maternal deaths a year, while it irreparably limits the cognitive development of children and leads to poor outcomes in pregnancy.

Currently available iron compounds are cheap and readily available, but constitute a non-physiological approach to providing iron that leads to significant side effects. Consequently, iron deficiency and IDA remain without an effective treatment, particularly in populations with high burden of infectious diseases. So far, despite considerable investment in the past 25 years in nutrition interventions with iron supplementation and fortification, we have been unable to significantly decrease the burden of this disease in resource-poor countries.

If we are to eliminate this condition in the future, it is imperative to look beyond the strategies used until now and we should make an effort to combine community engagement and social science approaches to optimise supplementation and fortification programmes.

Type
Conference on ‘Nutrient–nutrient interaction’
Copyright
Copyright © The Authors 2018 

Iron deficiency anaemia: the scale of the problem

Iron deficiency remains the most prevalent nutritional deficiency worldwide, affecting an estimated 4–6 billion people. Iron deficiency is also the main cause of anaemia in children and women in both high or low infection burden settings(Reference Wirth, Woodruff and Engle-Stone1, Reference Engle-Stone, Aaron and Huang2). Two billion people, about 30 % of the world's population, and 43 % of children between the age 6–59 months, are anaemic and the prevalence of anaemia is five times higher in low- and middle-income countries (LMIC) than in high-income countries(3, Reference Miller4). As such, iron deficiency anaemia (IDA) is the largest nutritional deficiency disorder in the world and one of the five leading causes of global disease burden(3). At any given moment, more individuals suffer from IDA than any other health problem with a staggering 1·24 billion affected individuals worldwide(3) (see Fig. 1).

Fig. 1. Prevalence of iron deficiency, iron deficiency anaemia (IDA) and anaemia worldwide. Adapted from(3).

Iron deficiency is associated with multiple pathologies, including anaemia and defective organ function and formation(Reference Prentice, Mendoza and Pereira5). Iron deficiency in the absence of anaemia is somewhat subtler in its manifestations than other micronutrient deficiencies, despite being a major contributor to ill health, premature death and lost earnings in developing countries. Even mild iron deficiency appears to impair intellectual development in young children and is lowering national intelligence quotients(Reference Grantham-McGregor and Ani6Reference Halterman, Kaczorowski and Aligne10), while overt IDA is associated with an increased risk of serious morbidity, poor motor and mental development in children, reduced work capacity in adults, poor pregnancy outcomes and impaired immunity: all of which make it one of the most expensive diseases in the world according to the WHO, causing an estimated 4·05 % of lost gross domestic product globally(Reference Horton and Ross11). For WHO, reducing the prevalence of IDA is one of its six priorities, and it is estimated that appropriate treatment of IDA not only could restore individual's health, but also raise national productivity levels by as much as 20 %(Reference Horton and Ross1114).

Nevertheless, in the past 25 years, not much has changed in LMIC, where IDA maintains its top position as the leading cause of years lived with disability(3), and is responsible for over 50 000 deaths annually(Reference Wang, Naghavi and Allen15), with iron deficiency contributing to 120 000 maternal deaths per year(16). This is perhaps not surprising, since eliminating IDA in these countries is particularly challenging due to the double burden of IDA and infectious diseases, and malaria, HIV/AIDS, tuberculosis, hookworm and other intestinal parasitic or bacterial infections, all contribute to the anaemia burden.

Iron deficiency in the UK

Recently, Public Health England and the Food Standards Agency has published the combined results from years 7 and 8 (2014/2015–2015/2016) of the National Diet and Nutrition Survey(17), providing an update on results from years 5 and 6 published in 2016. Data on dietary iron intake as well as on biochemical body iron status indicated an increased risk of iron deficiency in girls aged 11–18 years and women aged 19–64 years. When the dietary iron intake was compared with government recommendations, the National Diet and Nutrition Survey reported that 54 % of girls and 27 % of women had iron intakes below the lower reference nutrient intake, which is considered inadequate for most individuals, as it represents the level of intake that is likely to be sufficient to meet the needs of only 2·5 % of the population. Interestingly, the proportion of children aged 1·5–3 years with dietary iron intake below the lower reference nutrient intake was greater than in the previous report (10 % compared with 6 % in the past)(17, Reference Geissler and Singh18).

Analysis of blood samples of 704 adults and 329 children in the same UK survey provided evidence of IDA (as indicated by low Hb levels) and low iron stores (plasma ferritin) in all age/sex groups in the population, with a higher prevalence in females. Based on the WHO criteria for the definition of iron deficiency and anaemia, the prevalence of IDA in the UK was 9 % for girls aged 11–18 years (v. 5 % in the previous report), while low iron stores were evident in 24 % of adolescent girls, 5 % of adult women and 1 % of older women(17). It is important to note that pregnant or breast-feeding women who may have different requirements were among the population groups that were excluded from the survey.

These data support that iron deficiency, and IDA, remain impactful nutritional deficiency disorders in the UK, despite the wide-ranging availability of fortified foods and iron supplements.

Strategies to prevent and treat iron deficiency

The most commonly used strategies to control iron deficiency are supplementation (including multi-micronutrient powders), food fortification, dietary diversification and control of parasitic and other infections. Supplementation or fortification programmes are often the selected short-term approaches because they are cost-effective and relatively easy to implement.

The current WHO recommendation for the prevention of IDA is for daily iron supplementation for 3 consecutive months in a year to all pre-menopausal women, adolescent girls and young children in countries with over 40 % anaemia prevalence (i.e. the majority of countries in sub-Saharan Africa and South East Asia) and intermitted supplementation in settings with lower prevalence(19, 20).

Iron compounds are widely available but typically generate a non-physiological bolus of bio-accessible and reactive ionic iron that can cause significant adverse effects, either in the colon (i.e. unabsorbed iron fraction) or in circulation (i.e. absorbed iron fraction)(Reference Sazawal, Black and Ramsan21Reference Prentice24). Meta-analysis of trials involving nearly 10 000 young children in developing countries have consistently shown that conventional soluble oral iron supplementation used to treat IDA is associated with increased infection including bloody diarrhoea(Reference Soofi, Cousens and Iqbal22, Reference Mayo-Wilson, Imdad and Junior25Reference Jaeggi, Kortman and Moretti27) and detrimental changes to the gut microbiome and gut inflammation(Reference Jaeggi, Kortman and Moretti27, Reference Zimmermann, Chassard and Rohner28), further increasing the burden from enteric infection and environmental enteropathy (i.e. persistent gut damage and inflammation that leads to malabsorption), which is a major cause of growth failure in children in resource-poor environments(Reference Naylor, Lu and Haque29, Reference Lin, Arnold and Afreen30). Wide-scale home fortification programmes hold promise for reaching at-risk populations with the use of multi-micronutrient powders(Reference De-Regil, Suchdev and Vist31). Nonetheless, progress has been painstakingly slow, in spite of the dimension of the problem, and strategies to control iron deficiency have failed to decrease the global burden of this deficiency. This is particularly the case for young children in sub-Saharan Africa, where iron supplementation has consistently shown limited efficacy and a potential for increasing infection risk(Reference Sazawal, Black and Ramsan21, Reference Zlotkin, Newton and Aimone26, Reference Mwangi, Phiri and Abkari32).

Our biggest challenge: iron bioavailability in high-infection settings

Infection

Many LMIC have high prevalence of malnutrition and infection, both of which contribute to a vicious cycle(Reference Muller and Krawinkel33). First, malnutrition weakens barrier and immune functions, allowing pathogens easier access and overall impairing the host's ability to fight pathogens(Reference Calder and Jackson34). Secondly, damage to the intestinal mucosa lining caused by enteric infection and inflammation will decrease nutrient status due to impaired absorption and diarrhoea; therefore predisposing to further infection and worsening nutritional status, and perpetuating a cycle of chronic infection and malnutrition(Reference Calder and Jackson34). This cycle of infection, malnutrition and persistent inflammation will lead to malabsorption and in the long term to growth failure and mortality(Reference Calder and Jackson34Reference Moore, Cole and Poskitt36).

The state of chronic infection leads to high levels of hepcidin and anaemia. Elevated hepcidin can decrease iron absorption, irrespective of the iron source, and also leads to low circulating serum iron and transferrin saturation (hypoferraemia), through the mobilisation of circulating iron to the iron stores in macrophages and the liver(Reference Atkinson, Armitage and Khandwala37Reference Nemeth and Ganz40).

Conversely, iron supplementation regimens can cause the release of a high bolus of iron in the blood, which may transiently overwhelm transferrin's iron binding capacity and generate redox-reactive non-transferrin-bound iron, which could be available to extra-cellular pathogenic organisms causing infection(Reference Brittenham, Andersson and Egli41Reference van der, Marx and Grobbee45). Indeed, a committee chaperoned by the WHO concluded that these non-physiological high doses of highly absorbable iron may bypass the naturally evolved systems that safely chaperone iron and be the cause of the excess non-transferrin-bound iron, therefore causing the afore-mentioned harmful effects(46). This conclusion, supported by a recent study, has shown that 3 h after consumption of a standard iron supplement dose, human serum greatly supports enhanced rates of replication of pathogenic bacteria(Reference Cross, Bradbury and Fulford47).

Nutritional immunity

Iron is an essential nutrient to virtually all human pathogens, and the most virulent and invasive strains are those with multiple iron acquisition and utilisation mechanisms(Reference Subashchandrabose and Mobley48Reference Parkhill, Wren and Thomson51). Furthermore, iron can also regulate evolutionary transitions between commensal and pathogenic states in microbes(Reference Barber and Elde52). Nutritional immunity is an innate defence mechanism characterised by the sequestration of iron and other essential trace elements from the circulation to reduce iron availability and limit pathogen growth and virulence(Reference Calder and Jackson34, Reference Barber and Elde53, Reference Bullen, Rogers and Spalding54).

There is a constant competition for iron between the human host and invasive pathogens(Reference Drakesmith and Prentice39). Anaemia is protective against malaria and iron supplementation removes this protective effect(Reference Goheen, Wegmuller and Bah55, Reference Clark, Goheen and Fulford56). Iron supplementation may also increase the risk of respiratory infections and other systemic infections, due to iron's ability to increase virulence and multiplication of pathogens(Reference Cross, Bradbury and Fulford47, Reference Armitage, Stacey and Giannoulatou57, Reference McDermid, Hennig and van der Sande58). Our body has evolved very effective ways to intentionally reduce circulating iron when challenged by infection in an attempt to ‘starve’ the pathogen. Central to nutritional immunity are high-affinity iron transport proteins, such as transferrin and lactoferrin, which allow us to ‘starve’ pathogens by maintaining free iron atoms scarce at concentrations of 10−18 m(Reference Bullen, Rogers and Spalding54, Reference Bilitewski, Blodgett and Duhme-Klair59, Reference Weinberg60). As aforementioned, a key aspect of our response to infection is the more recently discovered hormone hepcidin, the master regulator of body iron homeostasis, which allows our body to reduce circulating iron by supressing intestinal iron absorption and decreasing its release from body iron stores. Any iron supplementation regimen that aims to be effective in a high-infection burden setting needs to acknowledge the close interaction between iron and infection and this key aspect of our natural immunity cannot be overlooked.

Hepcidin/inflammation

Our body has no means of excreting excess iron and the control of body iron levels occurs by regulation of iron absorption. Absorption of iron from the gastrointestinal tract is tightly regulated by the systemic need for iron through the action of the hormone hepcidin(Reference Drakesmith and Prentice39, Reference Ganz61, Reference Nemeth, Tuttle and Powelson62). Hepcidin responds to changes in body iron stores, tissue hypoxia and demand for iron, and it alters absorption accordingly. Hepcidin, as the master regulator of iron homeostasis, is the main inhibitor of iron export from cells (including enterocytes, hepatocytes and macrophages) into blood circulation(Reference Prentice, Mendoza and Pereira5). When hepcidin is elevated, it binds to ferroportin (the only cellular iron exporter known) and it causes its internalisation and degradation, so that cellular iron is locked inside cells and cannot (1) be absorbed into the bloodstream or (2) be released from body iron stores(Reference Ganz and Nemeth63, Reference Deschemin and Vaulont64).

In IDA, hepcidin levels tend to be low to allow iron absorption, but in the presence of inflammation or in iron-replete individuals, hepcidin levels are usually high(Reference Jaeggi, Moretti and Kvalsvig38, Reference Bregman, Morris and Koch65, Reference Prentice, Doherty and Abrams66). Anaemia of inflammation, also called anaemia of chronic disease, is characterised by high hepcidin, high iron stores and low circulation iron(Reference Pasricha, Atkinson and Armitage67).

In anaemia of inflammation, when hepcidin levels are high, if non-physiological bolus doses of highly absorbable iron supplements are initiated, this iron will be poorly absorbed, while unabsorbed iron will just be excreted in the stools(Reference Prentice, Doherty and Abrams66). Iron absorption can be as high as 50 % when hepcidin is switched off and is virtually zero when hepcidin is high(Reference Prentice, Doherty and Abrams66).

Microbiome

Low ‘free’ iron availability in the colon is a major modulator of our gut microbiome particularly in early life or disease. Both iron deficiency and iron supplementation have an impact on the gut microbiome(Reference Kortman, Dutilh and Maathuis68, Reference Pereira, Aslam and Frazer69). The establishment of an adult-like microbiome requires bio-accessible iron to maintain diversity(Reference Pereira, Aslam and Frazer69, Reference Timmerman, Rutten and Boekhorst70), but in certain population groups, iron supplementation could cause or contribute to microbiome dysbiosis (see Fig. 2).

Fig. 2. Model of microbiome dysbiosis caused by an excess of bio-accessible iron (Fe) in the colon.

Excess unabsorbed dietary iron through fortification or supplementation can modify the gut microbiota equilibrium and favour the growth of pathogenic strains over barrier strains. Iron is essential for the virulence and colonisation of most enteropathogenic strains(Reference Kortman, Mulder and Richters71Reference Jadhav, Hussain and Devi74). However, gut bacteria species considered most beneficial to the host, i.e. barrier function bacteria such as members of the Lactobacillacae and Bifidobacteriacae families, have either no or a very low requirement for iron(Reference Weinberg75, Reference Paganini, Uyoga and Kortman76). Indeed, studies in children have shown the detrimental effects of currently used soluble forms of iron on the gut microbiome, suggesting that iron supplements decrease the abundance of beneficial bacteria groups such as Lactobacillus and Bifidobacterium and increase the relative abundance of potential enteropathogenic bacteria belonging to the Enterobacteriaceae family, and also increase intestinal inflammation(Reference Jaeggi, Kortman and Moretti27, Reference Zimmermann, Chassard and Rohner28, Reference Paganini, Uyoga and Kortman76). This gut microbiome dysbiosis combined with mucosal inflammation can lead to reduced resistance to infection, an increase in the risk of diarrhoea and an increase in gut permeability leading to endotoxaemia(Reference Soofi, Cousens and Iqbal22, Reference Jaeggi, Kortman and Moretti27, Reference Paganini, Uyoga and Kortman76Reference Cani, Bibiloni and Knauf79). In this way, current iron supplementation could be contributing to the vicious cycle of infection, diarrhoea and eventually causing further anaemia. Use of prebiotics together with iron supplements may be able to alleviate part of the gut-related adverse effects of the iron supplements, but further studies are needed(Reference Paganini, Uyoga and Kortman76).

Future outlook: science and beyond

Micronutrient nutrition is a key aspect of the international development agenda and is integral to the UN sustainable development goals. Recognising the importance of micronutrient nutrition is also paramount in defining strategies to address the double burden of undernutrition and overnutrition and the double burden of malnutrition and infection, both widely prevalent across most countries in the world. The WHO recently recognised that a considerable investment in building public health nutrition capacity is required, particularly in LMIC, alongside a persistent effort in obtaining scientific evidence for the effectiveness of existing and new nutrition programmes and interventions(Reference Delisle, Shrimpton and Blaney80). Additionally, obtaining scientific evidence through clinical trials and nutrition intervention studies in ‘virgin’ rural settings in LMIC also carries the ethical responsibility to add social value, with community engagement and capacity building, which goes beyond the scientific discovery. Traditionally, research communication has been reserved for scientific publications and news pieces in research websites. However, this strategy creates a science isolation that is not effective in today's world. Therefore, we must train researchers not only in their ability to conduct nutrition science research but also in their ability to communicate science to the wider community and general public. Equally we must equip public health professionals, nurses and clinicians in LMIC in their ability to engage with the local communities as much as in their knowledge of how to manage the prevention and treatment of various forms of malnutrition. A practical example of how we can combine scientific discovery, capacity building and community engagement is outlined later with the IHAT-GUT iron supplementation clinical trial in The Gambia.

Case study: IHAT-GUT trial in The Gambia

Since iron delivered naturally through foods may be a safer alternative(Reference Prentice, Mendoza and Pereira5), there is a growing interest in developing novel nano iron compounds or delivery systems to use in fortification and supplementation(Reference Powell, Bruggraber and Faria81Reference Pisani, Riccio and Sabbatini84). One of the strategies proposed is to use a mimetic of the iron core of ferritin, namely iron hydroxide adipate tartrate (IHAT), to alleviate the adverse effects associated with reactive ionic iron(Reference Pereira, Bruggraber and Faria85). The hypothesis is that supplementation with IHAT will cause a slow release of iron into the blood, a lower rise of transferrin saturation, resembling that of natural food iron(Reference Pereira, Bruggraber and Faria85Reference Aslam, Frazer and Faria87), and may constitute a safer option for supplementation, particularly in population settings with high-infection burden. Pre-clinical data indicate that IHAT does not accumulate in the intestinal mucosa and can promote a beneficial microbiota after dysbiosis(Reference Pereira, Bruggraber and Faria85, Reference Aslam, Frazer and Faria87). IHAT and standard-of-care ferrous sulphate, are currently being tested in a randomised placebo-controlled double-blind clinical trial taking place at the MRC Unit The Gambia at the London School of Hygiene & Tropical Medicine. The IHAT-GUT trial (NCT02941081) is enrolling about 700 young children living in some of the poorest and more rural communities in The Gambia, where risk of infection and diarrhoea are very high.

As add-on projects to the IHAT-GUT trial, we have conducted nutrition training workshops and we are also conducting a social science project, where we document the ‘ins’ and ‘outs’ of setting up a nutrition intervention clinical trial, aimed at translating novel chemistry into public benefit, in a rural setting in The Gambia (see Fig. 3). We believe this social science aspect of research training is as important to global health and nutrition science researchers as the training in nutrition science itself. The social science project is gathering the personal views from clinical and field study staff, all of whom where based in The Gambia before the trial commenced, and will, hopefully, demonstrate that it is possible to translate international science to rural Africa and that it is possible to conduct nutrition clinical trials in these rural settings to ICH-GCP international standards. Likewise, it will illustrate the many challenges faced by the researchers, for example, how to implement clinical studies in new settings, how to attract motivated and skilled staff, etc. This project will help staff and communities to embrace these studies and to acknowledge the added value of being part of nutrition research. We envisage that such information will inspire the next generation of researchers in global health and international nutrition, from academia or industry, to translate more international science to rural resource-poor settings.

Fig. 3. Systems model integrating social value and capacity building in an iron supplementation trial conducted to ICH-GCP standards in The Gambia.

Conclusion

Anaemia is multifactorial as it is caused by both poor diet and high levels of infection and inflammation. Nowhere more than in sub-Saharan Africa do these causes coexist, and eliminating iron deficiency and anaemia in African children and women still remains one of the main nutritional challenges of the twenty-first century. Iron supplementation in malaria-endemic areas or in regions with high-infection burden needs to be implemented with caution, should target those who are anaemic or at a high risk of deficiency, and be accompanied by strategies to prevent and treat malaria, hookworm, schistosomiasis and other nutrient deficiencies, such as vitamin B12, folate and vitamin A(Reference Mwangi, Phiri and Abkari32, 46).

In spite of many successes, it is clear that more needs to be done to eliminate iron deficiency in LMIC. Today, public health interventions that can correct or prevent iron and other micronutrient deficiencies merit the highest priority for national and international organisations. Routine monitoring for compliance, research into the best delivery vehicles, ensuring the continuous availability of supplements or fortified foods at the community level, engaging local community and community leaders, and capacity building are the key factors for sustaining achievements and increasing the success of such programmes. Only then will we fully value the World Bank comment: ‘the control of vitamin and mineral deficiencies is one of the most extraordinary development-based scientific advances of recent years. Probably no other technology available today offers as large an opportunity to improve lives and accelerate development at such low cost in such short time.

Acknowledgements

The authors are very grateful to all staff and participants of the IHAT-GUT trial for their continued support and inspiration.

Financial Support

D. I. A. P. and the IHAT-GUT trial are supported by a Bill and Melinda Gates Foundation Grand Challenges New Interventions for Global Health award. Funding for the add-on social value projects was received from the Cambridge-Africa Alborada Research Fund and the University College London.

Conflicts of Interest

D. I. A. P. is one of the inventors of the IHAT iron supplementation technology for which she could receive future awards to inventors through the MRC Awards to Inventor scheme. Notwithstanding, the authors declare no conflict of interest.

Authorship

All authors contributed to the writing of this manuscript. The opinions expressed in this article are the authors’ own and do not reflect necessarily the view of their affiliated institutions.

References

1.Wirth, JP, Woodruff, BA, Engle-Stone, R et al. (2017) Predictors of anemia in women of reproductive age: Biomarkers Reflecting Inflammation and Nutritional Determinants of Anemia (BRINDA) project. Am J Clin Nutr 106, 416S427S.Google Scholar
2.Engle-Stone, R, Aaron, GJ, Huang, J et al. (2017) Predictors of anemia in preschool children: Biomarkers Reflecting Inflammation and Nutritional Determinants of Anemia (BRINDA) project. Am J Clin Nutr 106, 402S415S.Google Scholar
3.GBD 2016 Disease and Injury Incidence and Prevalence Collaborators (2017) Global, regional, and national incidence, prevalence, and years lived with disability for 328 diseases and injuries for 195 countries, 1990–2016: a systematic analysis for the Global Burden of Disease Study 2016. Lancet 390, 12111259.Google Scholar
4.Miller, JL (2013) Iron deficiency anemia: a common and curable disease. Cold Spring Harb Perspect Med 3, a011866.Google Scholar
5.Prentice, AM, Mendoza, YA, Pereira, D et al. (2017) Dietary strategies for improving iron status: balancing safety and efficacy. Nutr Rev 75, 4960.Google Scholar
6.Grantham-McGregor, S & Ani, C (2001) A review of studies on the effect of iron deficiency on cognitive development in children. J Nutr 131, 649S666S; discussion 666S–668S.Google Scholar
7.Pollitt, E, Saco-Pollitt, C, Leibel, RL et al. (1986) Iron deficiency and behavioral development in infants and preschool children. Am J Clin Nutr 43, 555565.Google Scholar
8.McCann, JC & Ames, BN (2007) An overview of evidence for a causal relation between iron deficiency during development and deficits in cognitive or behavioral function. Am J Clin Nutr 85, 931945.Google Scholar
9.Lozoff, B, Jimenez, E & Smith, JB (2006) Double burden of iron deficiency in infancy and low socioeconomic status: a longitudinal analysis of cognitive test scores to age 19 years. Arch Pediatr Adolesc Med 160, 11081113.Google Scholar
10.Halterman, JS, Kaczorowski, JM, Aligne, CA et al. (2001) Iron deficiency and cognitive achievement among school-aged children and adolescents in the United States. Pediatrics 107, 13811386.Google Scholar
11.Horton, S & Ross, J (2007) The economics of iron deficiency (vol 28, pg 51, 2003). Food Policy 32, 141143.Google Scholar
12.McGuire, S (2015) World Health Organization. Comprehensive implementation plan on maternal, infant, and young child nutrition. Geneva, Switzerland, 2014. Adv Nutr 6, 134135.Google Scholar
13.Haas, JD & Brownlie, TT (2001) Iron deficiency and reduced work capacity: a critical review of the research to determine a causal relationship. J Nutr 131, 676S688S; discussion 688S–690S.Google Scholar
14.WHO (2014) Global Nutrition Targets 2025: Anaemia Policy Brief (WHO/NMH/NHD/14·4). Geneva: World Health Organization.Google Scholar
15.Wang, H, Naghavi, M, Allen, C et al. (2017) Global, regional, and national life expectancy, all-cause mortality, and cause-specific mortality for 249 causes of death, 1980–2015: a systematic analysis for the Global Burden of Disease Study 2015. Lancet 388, 14591544.Google Scholar
16.WHO (2009) Global Health Risks. Mortality and Burden of Disease Attributable to Selected Major Risks. Geneva: World Health Organization.Google Scholar
17.Public Health England and Food Standards Agency (2018) National Diet and Nutrition Survey. Results from Years 7–8 (combined) of the Rolling Programme (2014/15 to 2015/16). London: Public Health England.Google Scholar
18.Geissler, C & Singh, M (2011) Iron, meat and health. Nutrients 3, 283316.Google Scholar
19.WHO (2016) Guideline: Daily Iron Supplementation in Infants and Children. Geneva, Switzerland: World Health Organization.Google Scholar
20.WHO (2016) Guideline: Daily Iron Supplementation in Adult Women and Adolescent Girls. Geneva, Switzerland: World Health Organization.Google Scholar
21.Sazawal, S, Black, RE, Ramsan, M et al. (2006) Effects of routine prophylactic supplementation with iron and folic acid on admission to hospital and mortality in preschool children in a high malaria transmission setting: community-based, randomised, placebo-controlled trial. Lancet 367, 133143.Google Scholar
22.Soofi, S, Cousens, S, Iqbal, SP et al. (2013) Effect of provision of daily zinc and iron with several micronutrients on growth and morbidity among young children in Pakistan: a cluster-randomised trial. Lancet 382, 2940.Google Scholar
23.Prentice, AM, Verhoef, H & Cerami, C (2013) Iron fortification and malaria risk in children. JAMA 310, 914915.Google Scholar
24.Prentice, AM (2008) Iron metabolism, malaria, and other infections: what is all the fuss about? J Nutr 138, 25372541.Google Scholar
25.Mayo-Wilson, E, Imdad, A, Junior, J et al. (2014) Preventive zinc supplementation for children, and the effect of additional iron: a systematic review and meta-analysis. BMJ Open 4, e004647.Google Scholar
26.Zlotkin, S, Newton, S, Aimone, AM et al. (2013) Effect of iron fortification on malaria incidence in infants and young children in Ghana: a randomized trial. JAMA 310, 938947.Google Scholar
27.Jaeggi, T, Kortman, GA, Moretti, D et al. (2015) Iron fortification adversely affects the gut microbiome, increases pathogen abundance and induces intestinal inflammation in Kenyan infants. Gut 64, 731742.Google Scholar
28.Zimmermann, MB, Chassard, C, Rohner, F et al. (2010) The effects of iron fortification on the gut microbiota in African children: a randomized controlled trial in Cote d'Ivoire. Am J Clin Nutr 92, 14061415.Google Scholar
29.Naylor, C, Lu, M, Haque, R et al. (2015) Environmental enteropathy, oral vaccine failure and growth faltering in infants in Bangladesh. EBioMedicine 2, 17591766.Google Scholar
30.Lin, A, Arnold, BF, Afreen, S et al. (2013) Household environmental conditions are associated with enteropathy and impaired growth in rural Bangladesh. Am J Trop Med Hyg 89, 130137.Google Scholar
31.De-Regil, LM, Suchdev, PS, Vist, GE et al. (2013) Home fortification of foods with multiple micronutrient powders for health and nutrition in children under two years of age (review). Evid Based Child Health Cochrane Rev J 8, 112201.Google Scholar
32.Mwangi, MN, Phiri, KS, Abkari, A et al. (2017) Iron for Africa-Report of an expert workshop. Nutrients 9, E576.Google Scholar
33.Muller, O & Krawinkel, M (2005) Malnutrition and health in developing countries. CMAJ 173, 279286.Google Scholar
34.Calder, PC & Jackson, AA (2000) Undernutrition, infection and immune function. Nutr Res Rev 13, 329.Google Scholar
35.Deen, JL, Walraven, GE & von Seidlein, L (2002) Increased risk for malaria in chronically malnourished children under 5 years of age in rural Gambia. J Trop Pediatr 48, 7883.Google Scholar
36.Moore, SE, Cole, TJ, Poskitt, EM et al. (1997) Season of birth predicts mortality in rural Gambia. Nature 388, 434.Google Scholar
37.Atkinson, SH, Armitage, AE, Khandwala, S et al. (2014) Combinatorial effects of malaria season, iron deficiency, and inflammation determine plasma hepcidin concentration in African children. Blood 123, 32213229.Google Scholar
38.Jaeggi, T, Moretti, D, Kvalsvig, J et al. (2013) Iron status and systemic inflammation, but not gut inflammation, strongly predict gender-specific concentrations of serum hepcidin in Infants in Rural Kenya. PLoS ONE 8, e57513.Google Scholar
39.Drakesmith, H & Prentice, AM (2012) Hepcidin and the iron-infection axis. Science 338, 768772.Google Scholar
40.Nemeth, E & Ganz, T (2014) Anemia of inflammation. Hematol Oncol Clin North Am 28, 671681, vi.Google Scholar
41.Brittenham, GM, Andersson, M, Egli, I et al. (2014) Circulating non-transferrin-bound iron after oral administration of supplemental and fortification doses of iron to healthy women: a randomized study. Am J Clin Nutr 100, 813820.Google Scholar
42.Hutchinson, C, Al-Ashgar, W, Liu, DY et al. (2004) Oral ferrous sulphate leads to a marked increase in pro-oxidant nontransferrin-bound iron. Eur J Clin Invest 34, 782784.Google Scholar
43.Schumann, K, Solomons, NW, Orozco, M et al. (2013) Differences in circulating non-transferrin-bound iron after oral administration of ferrous sulfate, sodium iron EDTA, or iron polymaltose in women with marginal iron stores. Food Nutr Bull 34, 185193.Google Scholar
44.Brissot, P, Ropert, M, Le Lan, C et al. (2012) Non-transferrin bound iron: a key role in iron overload and iron toxicity. Biochim Biophys Acta 1820, 403410.Google Scholar
45.van der, A Dl, Marx, JJ, Grobbee, DE et al. (2006) Non-transferrin-bound iron and risk of coronary heart disease in postmenopausal women. Circulation 113, 19421949.Google Scholar
46.WHO (2007) Conclusions and recommendations of the WHO Consultation on prevention and control of iron deficiency in infants and young children in malaria-endemic areas. Food Nutr Bull 28, S621S627.Google Scholar
47.Cross, JH, Bradbury, RS, Fulford, AJ et al. (2015) Oral iron acutely elevates bacterial growth in human serum. Sci Rep 5, 16670.Google Scholar
48.Subashchandrabose, S & Mobley, HLT (2015) Virulence and fitness determinants of uropathogenic Escherichia coli. Microbiol Spectr 3, UTI-0015-2012.Google Scholar
49.Okeke, IN, Scaletsky, ICA, Soars, EH et al. (2004) Molecular epidemiology of the iron utilization genes of enteroaggregative Escherichia coli. J Clin Microbiol 42, 3644.Google Scholar
50.Connolly, JPR, Finlay, BB & Roe, AJ (2015) From ingestion to colonization: the influence of the host environment on regulation of the LEE encoded type III secretion system in enterohaemorrhagic Escherichia coli. Front Microbiol 6, 568.Google Scholar
51.Parkhill, J, Wren, BW, Thomson, NR et al. (2001) Genome sequence of Yersinia pestis, the causative agent of plague. Nature 413, 523527.Google Scholar
52.Barber, MF & Elde, NC (2015) Buried treasure: evolutionary perspectives on microbial iron piracy. Trends Genet 31, 627636.Google Scholar
53.Barber, MF & Elde, NC (2014) Escape from bacterial iron piracy through rapid evolution of transferrin. Science 346, 1362.Google Scholar
54.Bullen, JJ, Rogers, HJ, Spalding, PB et al. (2006) Natural resistance, iron and infection: a challenge for clinical medicine. J Med Microbiol 55, 251258.Google Scholar
55.Goheen, MM, Wegmuller, R, Bah, A et al. (2016) Anemia offers stronger protection than sickle cell trait against the erythrocytic stage of falciparum malaria and this protection is reversed by iron supplementation. EBioMedicine 14, 123130.Google Scholar
56.Clark, MA, Goheen, MM, Fulford, A et al. (2014) Host iron status and iron supplementation mediate susceptibility to erythrocytic stage Plasmodium falciparum. Nat Commun 5, 4446.Google Scholar
57.Armitage, AE, Stacey, AR, Giannoulatou, E et al. (2014) Distinct patterns of hepcidin and iron regulation during HIV-1, HBV, and HCV infections. Proc Natl Acad Sci USA 111, 1218712192.Google Scholar
58.McDermid, JM, Hennig, BJ, van der Sande, M et al. (2013) Host iron redistribution as a risk factor for incident tuberculosis in HIV infection: an 11-year retrospective cohort study. BMC Infect Dis 13, 48.Google Scholar
59.Bilitewski, U, Blodgett, JAV, Duhme-Klair, AK et al. (2017) Chemical and biological aspects of nutritional immunity-perspectives for new anti-infectives that target iron uptake systems. Angew Chem Int Ed Engl 56, 1436014382.Google Scholar
60.Weinberg, ED (1975) Nutritional immunity. Host's attempt to withhold iron from microbial invaders. JAMA 231, 3941.Google Scholar
61.Ganz, T (2003) Hepcidin, a key regulator of iron metabolism and mediator of anemia of inflammation. Blood 102, 783788.Google Scholar
62.Nemeth, E, Tuttle, MS, Powelson, J et al. (2004) Hepcidin regulates cellular iron efflux by binding to ferroportin and inducing its internalization. Science 306, 20902093.Google Scholar
63.Ganz, T & Nemeth, E (2012) Hepcidin and iron homeostasis. Biochim Biophys Acta 1823, 14341443.Google Scholar
64.Deschemin, JC & Vaulont, S (2013) Role of hepcidin in the setting of hypoferremia during acute inflammation. PLoS ONE 8, e61050.Google Scholar
65.Bregman, DB, Morris, D, Koch, TA et al. (2013) Hepcidin levels predict nonresponsiveness to oral iron therapy in patients with iron deficiency anemia. Am J Hematol 88, 97101.Google Scholar
66.Prentice, AM, Doherty, CP, Abrams, SA et al. (2012) Hepcidin is the major predictor of erythrocyte iron incorporation in anemic African children. Blood 119, 19221928.Google Scholar
67.Pasricha, SR, Atkinson, SH, Armitage, AE et al. (2014) Expression of the iron hormone hepcidin distinguishes different types of anemia in African children. Sci Transl Med 6, 235re233.Google Scholar
68.Kortman, GA, Dutilh, BE, Maathuis, AJ et al. (2015) Microbial metabolism shifts towards an adverse profile with supplementary iron in the TIM-2 In vitro model of the human colon. Front Microbiol 6, 1481.Google Scholar
69.Pereira, DIA, Aslam, MF, Frazer, DM et al. (2015) Dietary iron depletion at weaning imprints low microbiome diversity and this is not recovered with oral nano Fe(III). Microbiologyopen 4, 1227.Google Scholar
70.Timmerman, HM, Rutten, N, Boekhorst, J et al. (2017) Intestinal colonisation patterns in breastfed and formula-fed infants during the first 12 weeks of life reveal sequential microbiota signatures. Sci Rep 7, 8327.Google Scholar
71.Kortman, GA, Mulder, ML, Richters, TJ et al. (2015) Low dietary iron intake restrains the intestinal inflammatory response and pathology of enteric infection by food-borne bacterial pathogens. Eur J Immunol 45, 25532567.Google Scholar
72.Kortman, GA, Roelofs, RW, Swinkels, DW et al. (2014) Iron-induced virulence of Salmonella enterica serovar typhimurium at the intestinal epithelial interface can be suppressed by carvacrol. Antimicrob Agents Chemother 58, 16641670.Google Scholar
73.Kortman, GA, Boleij, A, Swinkels, DW et al. (2012) Iron availability increases the pathogenic potential of Salmonella typhimurium and other enteric pathogens at the intestinal epithelial interface. PLoS ONE 7, e29968.Google Scholar
74.Jadhav, S, Hussain, A, Devi, S et al. (2011) Virulence characteristics and genetic affinities of multiple drug resistant uropathogenic Escherichia coli from a semi urban locality in India. PLoS ONE 6, e18063.Google Scholar
75.Weinberg, ED (1997) The lactobacillus anomaly: total iron abstinence. Perspect Biol Med 40, 578583.Google Scholar
76.Paganini, D, Uyoga, MA, Kortman, GAM et al. (2017) Prebiotic galacto-oligosaccharides mitigate the adverse effects of iron fortification on the gut microbiome: a randomised controlled study in Kenyan infants. Gut 66, 19561967.Google Scholar
77.Mullen, A, Gosset, L, Larke, N et al. (2012) The effects of micronutrient-fortified complementary/replacement food on intestinal permeability and systemic markers of inflammation among maternally HIV-exposed and unexposed Zambian infants. Br J Nutr 107, 893902.Google Scholar
78.Brunser, O, Espinoza, J, Araya, M et al. (1993) Chronic iron intake and diarrhoeal disease in infants. A field study in a less-developed country. Eur J Clin Nutr 47, 317326.Google Scholar
79.Cani, PD, Bibiloni, R, Knauf, C et al. (2008) Changes in gut microbiota control metabolic endotoxemia-induced inflammation in high-fat diet-induced obesity and diabetes in mice. Diabetes 57, 14701481.Google Scholar
80.Delisle, H, Shrimpton, R, Blaney, S et al. (2017) Capacity-building for a strong public health nutrition workforce in low-resource countries. Bull World Health Organ 95, 385388.Google Scholar
81.Powell, JJ, Bruggraber, SFA, Faria, N et al. (2014) A nano-disperse ferritin-core mimetic that efficiently corrects anemia without luminal iron redox activity. Nanomed Nanotechnol Biol Med 10, 15291538.Google Scholar
82.von Moos, LM, Schneider, M, Hilty, FM et al. (2017) Iron phosphate nanoparticles for food fortification: biological effects in rats and human cell lines. Nanotoxicology 11, 496506.Google Scholar
83.Hilty, FM, Arnold, M, Hilbe, M et al. (2010) Iron from nanocompounds containing iron and zinc is highly bioavailable in rats without tissue accumulation. Nat Nanotechnol 5, 374380.Google Scholar
84.Pisani, A, Riccio, E, Sabbatini, M et al. (2015) Effect of oral liposomal iron versus intravenous iron for treatment of iron deficiency anaemia in CKD patients: a randomized trial. Nephrol Dial Transplant 30, 645652.Google Scholar
85.Pereira, DIA, Bruggraber, SFA, Faria, N et al. (2014) Nanoparticulate iron(III) oxo-hydroxide delivers safe iron that is well absorbed and utilised in humans. Nanomed Nanotechnol Biol Med 10, 18771886.Google Scholar
86.Pereira, DI, Mergler, BI, Faria, N et al. (2013) Caco-2 cell acquisition of dietary iron(III) invokes a nanoparticulate endocytic pathway. PLoS ONE 8, e81250.Google Scholar
87.Aslam, MF, Frazer, DM, Faria, N et al. (2014) Ferroportin mediates the intestinal absorption of iron from a nanoparticulate ferritin core mimetic in mice. FASEB J 28, 36713678.Google Scholar
Figure 0

Fig. 1. Prevalence of iron deficiency, iron deficiency anaemia (IDA) and anaemia worldwide. Adapted from(3).

Figure 1

Fig. 2. Model of microbiome dysbiosis caused by an excess of bio-accessible iron (Fe) in the colon.

Figure 2

Fig. 3. Systems model integrating social value and capacity building in an iron supplementation trial conducted to ICH-GCP standards in The Gambia.