1.Introduction
The year 2015 witnessed perhaps the greatest human migration ever seen from the underdeveloped, low socioeconomic countries of the Middle East, North Africa and Central and South America. These immigrants were heading to developed northern European countries and North America. Australia also has a history of welcoming immigrants and more recently Middle Eastern refugees fleeing war zones. However, the medical history of these migrant populations, including prenatal and early life exposures, may present unforeseen health problems requiring prophylactic measures.
It is likely that, if not addressed, the health problems caused by prenatal and early life under nutrition and starvation may result in medical conditions that place additional burden on the health system over the coming decades.
2. Historical Evidence
Scientific findings regarding immediate effects of starvation on humans were published in 1941. These unique studies were conducted by 28 doctors incarcerated in the Warsaw ghetto during WWII. The research was performed over 17 months and documented in detail the effects of starvation on people who were given daily rations consisting of less than 800 calories and only 20-30 g protein. By July 1942, more than 40,000 people had died
[1, 2].
The doctors observed multiple effects of starvation, particularly on the children, and most notably on their metabolic and endocrine functions. Their bone histology showed signs of "Hunger Osteopathy", including both architectural and matrix changes, namely osteoporosis and osteomalacia. Bone healing did not occur, making surgical fixation inexpedient
[1].
While the delayed effect of malnutrition on human development has been established for some time, rigorous scientific studies on the epidemiology of diseases associated with starvation, have only been conducted more recently
[3-5]. The exact effect of early-life malnutrition on the development of adult diseases was discovered in the late 20th century, on populations exposed to hunger during their embryonic, foetal, neonatal or childhood periods, some 40-50 years earlier
[6-7].
During WWII, starvation was used by the German authorities as a weapon for submission or punishment, for example in Leningrad and the Netherlands. Starvation also occurred on the Channel Islands in 1944 when food supply was cut off as a result of the Allied invasion of Normandy.
During the Leningrad siege the German army surrounded the city for over two years (September 1941 to January 1944) and the population of 2.9 million people, including 0.5 million children, were cut-off from food supplies. Approximately 830,000 people lost their lives
[8]. Forty years later, studies were conducted at the Karolinska Institute, Sweden, investigating the effects on children born to mothers who survived the siege with sustenance of only 300-800 calories per day during pregnancy. Findings showed a direct relationship between gestational deprivation and cardiovascular disease
[8,9].
In the Netherlands in November 1944, as a reprisal for interference with essential railway lines, a severe food embargo (‘the Dutch embargo’) was instituted by the occupying German forces upon the densely populated Western Netherlands. The food supplied per person was gradually reduced from an initial 1000 calories per day, down to 400 calories a day by April 1945. In total, 18,000 deaths were recorded as directly related to the famine. In a study conducted at Amsterdam University, the health records of those born during the siege found that they were small in stature, had an increased risk of metabolic disease and obesity, and had an increased mortality rate by age 50, compared to controls from the Eastern Netherlands. Changes were also identified in reproductive function, including early menopause, and increased incidence of breast and colon malignancy
[10-12].
The significant difference between the Leningrad siege and the Dutch embargo was that after liberation, the ‘catch up’ feeding in the Dutch cohort was rapid and sudden, whilst in Leningrad, the economic recovery was slow and occurred over a period of several years.
The famine resulting from the German occupation of the Channel Islands (1940-1945) involved the entire population of 60,000 people as well as the occupiers who had been left behind as the invasion of Normandy in 1944 by-passed the islands. The food supply was cut off, and by the end of 1944/early 1945 the shortage had become critical
[13-14]. It was found that children exposed to under-nutrition had increased risk of cardiovascular morbidity, delayed menarche, and a trend towards increased risk of breast cancer compared to the matched controls evacuated to England.
Further studies using birth registry data on a large peacetime population of England and Wales added to evidence for the effect of early-life nutrition on the development of diseases in adults
[5-18].
This historical data led to the well-known ‘Barker Hypothesis’ which proposes that there is adjustment in utero via a "re-printing" or "programming" mechanism. In 1992 further hypotheses that metabolic syndrome may result from in utero starvation and developmental plasticity were published
[4,5].
In more recent studies, a predisposition to the development of metabolic disease in the Nauru population, was observed once the island became commercially enriched. The mineral boom in the country led to "westernisation" of diet, and the Micronesian islanders were at increased risk of obesity and diabetes, and was trans-generational
[6].
Similarly, a Swedish study on the Biafran population, approximately 40 years after exposure to starvation, also observed increased occurrence of metabolic syndrome
[7].Most strikingly, however, was the fact that the theory that exposure to early life starvation increasing the risk of metabolic syndrome has been reproduced experimentally in animal models in numerous experiments.
However, none of the above initial studies dealt with bone mineral metabolism. The fetal programming of musculo-skeletal pathology (adipose tissue, sarcopaenia, osteopaenia) and effect on health in later life was developed somewhat later.
There is a plethora of publications that support the theory that early life malnutrition and later metabolic changes may be influential in programming increased risk for adult osteoporosis.
Cooper et al.(2000) described the pathogenesis of osteoporosis and attributed a role to intrauterine programming. This was a confirmation of an earlier analysis of growth in infancy and bone mass in adult life. In 2004, osteoporosis was considered to be of developmental origin, with significant impact on the risk of future fracture
[18-25]. By 2012, the epigenetic influence on the developmental origin of osteoporosis was established
[26]. The possible generational transfer of the metabolic insult suffered by one generation is also of possible relevance.
Due to early life events and exposure to hardship, it is vital that countries accepting immigrants and refugees are prepared for the likely increased health risks. In particular, it is important that funds and resources are directed towards prevention of osteoporosis, including early diagnosis, immediate treatment and reduction of fracture and secondary fracture risk
[26-30] The cost-savings by taking a preventative approach are likely to be considerable.
The risks are three-fold: The primary risk is starvation in utero (whether total, partial or selective). The risks are also trimester sensitive, for example, protein and glucose deficiency may influence brain development in the first trimester; of cortisol and growth hormone deficiency may influence bone and internal organ development during the second trimester and limb and kidney development during the third trimester
[3,31].
Offering additional health screening of immigrant populations for markers such as Glycated Haemoglobin (HbA1c), lipids (triglycerides, cholesterol with fractions, and the screening for minerals (e.g., calcium, magnesium, phosphates, vitamin D), may be required. A minor cost, (totalling some AUS=32), would allow identification and instigation of preventative measures for those at risk of diabetes, of obesity, atherogenesis and bone demineralisation (our preferred term Low Nutrition osteopathy). Vitamin K deficiency may also lead to paradoxical mineralisation, namely epiphyseal and arterial wall calcifications, and premature osteoporosis in adults.
The secondary risk is sudden compensatory hyper alimentation, or so called “catch-up overfeeding”, leading to adolescent obesity, early adult glucose and lipid metabolic disturbance, and full-blown adult metabolic syndrome, including osteoporosis (high nutrition osteopathy). Studies have shown that catch-up bone metabolism is only partially achieved and may even exaggerate osteopathy by partial mineralisation
[25, 26].
An osteoporosis prevention program will require a significant educational effort; similar measures were taken in Austria following a nationwide review of diabetes in the population born during starvation periods
[3, 6, 7, 31]
The third risk is of the surprising findings (even if at times with limited statistical significance), on the frequency of malignancy in individuals starved during their intrauterine or perinatal life, compared to the rest of the population
[11,12,14,32]
3.Conclusion
Osteoporosis, just as with other metabolic disorders, was found to be programmed by conditions of early life, mostly intrauterine starvation or partial nutritional deprivation. It appeared in both genders, more prominently in females, with significant risk of increased fracture incidence. Nutritional science has advanced with major steps in the last quarter of a century, demonstrating two separate aspects of bone growth and presenting two separate but inter-dependent metabolic bone syndromes, both related to maternal starvation during pregnancy. Some children do not achieve catch-up growth, remain short into adulthood, with increased coronary and cerebral artery disease and osteoporosis
[33].