The quest for fitness may be detrimental to your long term mental state

We are often told how we should aim to have, and maintain, a healthy lifestyle. After all, being physically fit allows your body to function both in physical and mental aspects. Healthy body, healthy mind, right?

The only difficulty, if you can call it that, with exercise is that the first thing that we would normally consider is running, but it is not for everyone. Going forward for a certain distance or time has little meaning for some people, especially children.

The thing about running is that it has to have some appreciable meaning, so unless you have some derivative inner joy of measuring your progress using statistics, it is unlikely to hold your interest for the long term. A better form of exercise is though group sports, as the mental boredom of tracking fitness levels is negated in favour of the social dynamic.

Common group sports such as football  have a large following in England. The football season for example lasts from August to May and provides a welcome distraction during the cold winter months. It is also a simple game that can be improvised using other materials and played on all surfaces. No goalposts? Use bags or some other markers. No football? Use a tennis ball. It is often interesting to see children turn up at a field, establish the boundaries of play using trees and creates goalposts using caps or other loose materials and these are often sufficient for the game; at least until there is discussion about whether the “ball” hit the post or went in the goal after it flies over a set of keys intended to represent the goalpost.

There is increasing concern about the link between dementia and football. The pounding of the ball against a soft surface of the brain, when the ball is headed, over time can cause the destruction of cells and cell function. This is of particular concern in the case of children, whose brains and bodies are developing. This has been of significant interest as members of England’s 1966 World Cup winning squad have found to have developed dementia in their later years. Some of them cannot even remember being there in 1966!

It is not just the impact of ball on head that is concerning, but when the head is moved through a range of motion too quickly. Even though there is no impact on the head externally, internally there is damage as the brain is hitting the sides of the skull supposed to protect it.

It is not just football that we have to be concerned about. There is plenty of head and neck related impact in rugby and American football. In fact, in American football, the head related injuries for offensive and defensive linemen, who every forty seconds start a play by ramming into the player on the opposite side of the line,  and the list of dementia sufferers is growing continually. Some players have even sued the NFL for injuries suffered during the game.

Will the rules of football change so that heading the ball is banned? Don’t bet on it. That would change the fabric of the game so much as to ruin it. When the ball is swung in from a corner, what would you do if you couldn’t head it? The game will not change, but also don’t rule out a consortium of players in the future filing lawsuits for work-related injuries. Perhaps in the pursuit of fitness, it may be wiser to choose less impactful activities for the sake of long term health.

Why mental health problems will never go away

Many people will experience mental health difficulties at some point in their lives. As people go through life the demands on them increase, and over a prolonged period these can cause difficulty and ill health. These problems can manifest themselves both in mental and physical ways.

What kind of demands do people experience? One of these can be work-related. People may experience  stresses of looking for work, having to work in jobs which do not test their skills, or be involved in occupations  which require skills that are seemingly difficult to develop. Another common theme with adults that causes stress is having to work in a job which increasingly demands more of them, but does not remunerate them accordingly. In other words, they have to work more for less, and have to accept the gradual lowering of work conditions, but are unable to change jobs because they have already invested so much in it in terms of working years, but cannot leave and start afresh because the demands of a mortgage to pay off and a young family to provide for means they cannot start on a lower rung in a new occupation. Over a prolonged period, this can cause severe unhappiness.

Is it surprising that suicide affects men in their thirties and forties? This is a period for a man where work demands more, the mortgage needs paying, and the family demands more of his time and energy. It is unsurprising that having spent long periods in this sort of daily struggle, that men develop mental health problems which lead some to attempt suicide. But mental health does not just affect men. Among some of the this some women have to deal with are the struggles of bringing up children, the work life balance, the unfulfilled feel of not utilising their skills, and feeling isolated.

One of the ways ill health develops mentally is when people spend too long being pushed too hard for too long. Put under these kind of demands, the body shuts down as a self preservation measure. But the demands on the person don’t just go away. You may want a break from work. But this may not be possible or practical. In fact, the lack of an escape when you are aware you need one may be a greater trigger of mental illness, because it increases the feeling of being trapped.

It is little wonder that when people go through periods of mental ill health, an enforced period of short-term rest will allow them to reset their bearings to be able to continue at work, or return to work with some level of appropriate support. But this is only temporary.

With mental ill health problems, lifestyle adjustments need to be made for sufficient recovery.

Under the Equality Act (2010), your employer has a legal duty to make “reasonable adjustments” to your work.

Mental ill health sufferers could ask about working flexibly, job sharing, or a quiet room, a government report suggests.

The practicality of this however means more cost to the employer in having to make adjustments to accommodate the employee, and unless the employee is a valued one, whom the employer would like to keep, often the case is that they will be gradually phased out of the organisation.

In fact, when an employee attains a certain level of experience within an organisation, employers often ask more of them because they know these employees are locked in to their jobs, and have to accept these grudgingly, or risk losing their jobs, which they cannot do if they have dependents and financial commitments.
And you know the irony of it? The mental ill health sufferer already knows that. Which is why they don’t speak out for help in the first place.

If these employees complain, employers simply replace them with younger employees, who cost less, and who are willing to take on more responsibilities just to have a job. Any responsibilities the redundant employee had simply get divided up between his leftover colleagues, who are in turn asked to take on more responsibilities. They are next in line in the mental health illness queue.

And what if you are self employed? And have to work to support yourself and your dependents? The demands of the day to day are huge and don’t seem to go away.

You can see why mental health is  perceived a ticking time bomb. Organisations are not going to change to accommodate their employees because of cost, but keep pressing them to increase productivity without pay, knowing that they cannot say no, and when all the life and juice has been squeezed out of them, they can be chucked away and replaced with the next dispensable employee.

A ticking time bomb.

The financial considerations of investing in medicine and medical research

BBC News reports that a drug that would reduce the risk of HIV infection would result in cost savings of over £1bn over 80 years. Pre-exposure prophylaxis, or Prep, would reduce infection and hence lower the treatment costs for patients in the long term.

The catch? There is one. It’s the long term.

The cost of the treatment and prevention is such that its provision for the first twenty years – bundling together the cost of medical research and production of medicine – would result in a financial loss, and parity would only be achieved after a period of about thirty to forty years; this period is hard to define because it is dependent on what the drug would cost in the future.

Prep combines two anti-HIV drugs, emtricitabine and tenofovir. The medical trials behind it have concluded it has an effective rate of over one in five when it comes to protecting men who have unprotected sex with men from HIV infection. The exact figure is close to 86%.

Prep can be used either on a daily basis, or on what has been termed a sexual event basis – using it for two days before, during and after periods of unprotected sex.

The research model analysed the potential impact of Prep and found that it could reduce infection rates by over a quarter. The cost of the treatment itself, comparative to the cost of treating infection, would result in a saving over one billion pounds over eight years.

However, it does raise a few ethical questions. If the National Health Service is aiming to be a sustainable one – and one of the aims of sustainability is to empower citizens to take responsibility for their own health –  shouldn’t it be considering less about how it will balance the books, but spend more on education for prevention in the first place? The cost of producing Prep on the NHS would be £19.6 billion over 80 years; while the estimated savings from treatment would be £20.6 billion over the same period. Educating people not to have unprotected sex with those at the risk of HIV arguably would result in a higher saving over a lower time period. Perhaps the NHS should consider ways of reducing cost more significantly, rather than latching on to a cheaper prevention drug immediately. If consumer behaviour is not going to change, symptoms are still going to surface, and the provision of Prep on the NHS may only encourage less self-regulation and awareness.

Airbnb style recuperation for hospital patients

Would you welcome a stranger into your home? Would you have a spare room set aside for them? Perhaps not. But what if you were paid to do so? This is what some hospital bosses are considering to relieve overcrowding in hospital wards, that patients do their recuperating in private homes, rather than in the hospital. You offer a room if you have one available, and the hospital rents it from you for a patient. It is like an airbnb for hospitals.

On the face of it, this seems like a good idea. Hospital overcrowding is lessened, home owners get a bit of spare cash, the recuperating patient gets a bit of company … everyone’s happy. Patients staying out of hospitals mean that the backlog of operations can be cleared more quickly, resulting in a better streamlined NHS that benefits every citizen.

This idea is being piloted by the startup CareRooms. “Hosts”, who do not necessarily need to have previous experience in healthcare, could earn £50 a night and up to maximum of £1000 a month putting up local residents who are awaiting discharge from hospital. The pilot will start with 30 patients and the hope is that this will expand.

AgeUK claims that patients were being “marooned” in hospitals, taking up beds while 2.2 million days are lost annually to delayed transfers of care.

The specifics, however, do not seem to hold up to scrutiny. Who is responsible for the overall welfare of the patient? Once a patient is transferred to this “care” home, the responsibility of medical care is devolved to someone with basic first-aid training.

Prospective hosts are also required to heat up three microwave meals each day and supply drinks. Unfortunately it opens the issues of safeguarding, governance and possible financial and emotional abuse of people at their most vulnerable time.

The recuperating patients will “get access to a 24-hour call centre, tele-medical GP and promised GP consultation within four hours.”

The underlying question, though, is would you, though, want your loved ones to be put through this kind of care?

This is cost-cutting at its worst. The NHS is cutting costs, cutting ties and cutting responsibilities for those supposedly under its care. It would be a sad day if this kind of devolved responsibility plan became approved.

Media’s Marvellous Medicine

When it comes to our health, the media wields enormous influence over what we think. They tell us what’s good, what’s bad, what’s right and wrong, what we should and shouldn’t eat. When you think about it, that’s quite some responsibility. But do you really think that a sense of philanthropic duty is the driving force behind most of the health ‘news’ stories that you read? Who are we kidding? It’s all about sales, of course, and all too often that means the science plays second fiddle. Who wants boring old science getting in the way of a sensation-making headline?

When it comes to research – especially the parts we’re interested in, namely food, diet and nutrients – there’s a snag. The thing is, these matters are rarely, if ever, clear-cut. Let’s say there are findings from some new research that suggest a component of our diet is good for our health. Now academics and scientists are generally a pretty cautious bunch – they respect the limitations of their work and don’t stretch their conclusions beyond their actual findings. Not that you’ll think this when you hear about it in the media. News headlines are in your face and hard hitting. Fluffy uncertainties just won’t cut it. An attention-grabbing headline is mandatory; relevance to the research is optional. Throw in a few random quotes from experts – as the author Peter McWilliams stated, the problem with ‘experts’ is you can always find one ‘who will say something hopelessly hopeless about anything’ – and boom! You’ve got the formula for some seriously media-friendly scientific sex appeal, or as we prefer to call it, ‘textual garbage’. The reality is that a lot of the very good research into diet and health ends up lost in translation. Somewhere between its publication in a respected scientific journal and the moment it enters our brains via the media, the message gets a tweak here, a twist there and a dash of sensationalism thrown in for good measure, which leaves us floundering in a sea of half-truths and misinformation. Most of it should come with the warning: ‘does nothing like it says in the print’. Don’t get us wrong: we’re not just talking about newspapers and magazines here, the problem runs much deeper. Even the so-called nutrition ‘experts’, the health gurus who sell books by the millions, are implicated. We’re saturated in health misinformation.

Quite frankly, many of us are sick of this contagion of nutritional nonsense. So, before launching headlong into the rest of the book, take a step back and see how research is actually conducted, what it all means and what to watch out for when the media deliver their less-than-perfect messages. Get your head around these and you’ll probably be able to make more sense of nutritional research than most of our cherished health ‘gurus’.

Rule #1: Humans are different from cells in a test tube
At the very basic level, researchers use in-vitro testing, in which they isolate cells or tissues of interest and study them outside a living organism in a kind of ‘chemical soup’. This allows substances of interest (for example, a vitamin or a component of food) to be added to the soup to see what happens. So they might, for example, add vitamin C to some cancer cells and observe its effect. We’re stating the obvious now when we say that what happens here is NOT the same as what happens inside human beings. First, the substance is added directly to the cells, so they are often exposed to concentrations far higher than would normally be seen in the body. Second, humans are highly complex organisms, with intricately interwoven systems of almost infinite processes and reactions. What goes on within a few cells in a test tube or Petri dish is a far cry from what would happen in the body. This type of research is an important part of science, but scientists know its place in the pecking order – as an indispensable starting point of scientific research. It can give us valuable clues about how stuff works deep inside us, what we might call the mechanisms, before going on to be more rigorously tested in animals, and ultimately, humans. But that’s all it is, a starting point.

Rule #2: Humans are different from animals
The next logical step usually involves animal testing. Studying the effects of a dietary component in a living organism, not just a bunch of cells, is a big step closer to what might happen in humans. Mice are often used, due to convenience, consistency, a short lifespan, fast reproduction rates and a closely shared genome and biology to humans. In fact, some pretty amazing stuff has been shown in mice. We can manipulate a hormone and extend life by as much as 30%1. We can increase muscle mass by 60% in two weeks. And we have shown that certain mice can even regrow damaged tissues and organs.

So, can we achieve all of that in humans? The answer is a big ‘no’ (unless you happen to believe the X-Men are real). Animal testing might be a move up from test tubes in the credibility ratings, but it’s still a long stretch from what happens in humans. You’d be pretty foolish to make a lot of wild claims based on animal studies alone.

To prove that, all we need to do is take a look at pharmaceutical drugs. Vast sums of money (we’re talking hundreds of millions) are spent trying to get a single drug to market. But the success rate is low. Of all the drugs that pass in-vitro and animal testing to make it into human testing, only 11% will prove to be safe and effective enough to hit the shelves5. For cancer drugs the rate of success is only 5%5. In 2003, the President of Research and Development at pharmaceutical giant Pfizer, John La Mattina, stated that ‘only one in 25 early candidates survives to become a prescribed medicine’. You don’t need to be a betting person to see these are seriously slim odds.

Strip it down and we can say that this sort of pre-clinical testing never, ever, constitutes evidence that a substance is safe and effective. These are research tools to try and find the best candidates to improve our health, which can then be rigorously tested for efficacy in humans. Alas, the media and our nutrition gurus don’t appear to care too much for this. Taking research carried out in labs and extrapolating the results to humans sounds like a lot more fun. In fact, it’s the very stuff of many a hard-hitting newspaper headline and bestselling health book. To put all of this into context, let’s take just one example of a classic media misinterpretation, and you’ll see what we mean.

Rule #3: Treat headlines with scepticism
Haven’t you heard? The humble curry is right up there in the oncology arsenal – a culinary delight capable of curing the big ‘C’. At least that’s what the papers have been telling us. ‘The Spice Of Life! Curry Fights Cancer’ decreed the New York Daily News. ‘How curry can help keep cancer at bay’ and ‘Curry is a “cure for cancer”’ reported the Daily Mail and The Sun in the UK. Could we be witnessing the medical breakthrough of the decade? Best we take a closer look at the actual science behind the headlines.

The spice turmeric, which gives some Indian dishes a distinctive yellow colour, contains relatively large quantities of curcumin, which has purported benefit in Alzheimer’s disease, infections, liver disease, inflammatory conditions and cancer. Impressive stuff. But there’s a hitch when it comes to curcumin. It has what is known as ‘poor bioavailability’. What that means is, even if you take large doses of curcumin, only tiny amounts of it get into your body, and what does get in is got rid of quickly. From a curry, the amount absorbed is so miniscule that it is not even detectable in the body.

So what were those sensational headlines all about? If you had the time to track down the academic papers being referred to, you would see it was all early stage research. Two of the articles were actually referring to in-vitro studies (basically, tipping some curcumin onto cancer cells in a dish and seeing what effect it had).

Suffice to say, this is hardly the same as what happens when you eat a curry. The other article referred to an animal study, where mice with breast cancer were given a diet containing curcumin. Even allowing for the obvious differences between mice and humans, surely that was better evidence? The mice ate curcumin-containing food and absorbed enough for it to have a beneficial effect on their cancer. Sounds promising, until we see the mice had a diet that was 2% curcumin by weight. With the average person eating just over 2kg of food a day, 2% is a hefty 40g of curcumin. Then there’s the issue that the curcumin content of the average curry/turmeric powder used in curry is a mere 2%. Now, whoever’s out there conjuring up a curry containing 2kg of curry powder, please don’t invite us over for dinner anytime soon.

This isn’t a criticism of the science. Curcumin is a highly bio-active plant compound that could possibly be formulated into an effective medical treatment one day. This is exactly why these initial stages of research are being conducted. But take this basic stage science and start translating it into public health advice and you can easily come up with some far-fetched conclusions. Let us proffer our own equally absurd headline: ‘Curry is a Cause of Cancer’. Abiding by the same rules of reporting used by the media, we’ve taken the same type of in-vitro and animal-testing evidence and conjured up a completely different headline. We can do this because some studies of curcumin have found that it actually causes damage to our DNA, and in so doing could potentially induce cancer.

As well as this, concerns about diarrhoea, anaemia and interactions with drug-metabolizing enzymes have also been raised. You see how easy it is to pick the bits you want in order to make your headline? Unfortunately, the problem is much bigger than just curcumin. It could just as easily be resveratrol from red wine, omega-3 from flaxseeds, or any number of other components of foods you care to mention that make headline news.

It’s rare to pick up a newspaper or nutrition book without seeing some new ‘superfood’ or nutritional supplement being promoted on the basis of less than rigorous evidence. The net result of this shambles is that the real science gets sucked into the media vortex and spat out in a mishmash of dumbed-down soundbites, while the nutritional messages we really should be taking more seriously get lost in a kaleidoscope of pseudoscientific claptrap, peddled by a media with about as much authority to advise on health as the owner of the local pâtisserie.

Rule #4: Know the difference between association and causation
If nothing else, we hope we have shown that jumping to conclusions based on laboratory experiments is unscientific, and probably won’t benefit your long-term health. To acquire proof, we need to carry out research that involves actual humans, and this is where one of the greatest crimes against scientific research is committed in the name of a good story, or to sell a product.

A lot of nutritional research comes in the form of epidemiological studies. These involve looking at populations of people and observing how much disease they get and seeing if it can be linked to a risk factor (for example, smoking) or some protective factor (for example, eating fruit and veggies). And one of the most spectacular ways to manipulate the scientific literature is to blur the boundary between ‘association’ and ‘causation’. This might all sound very academic, but it’s actually pretty simple.

Confusing association with causation means you can easily arrive at the wrong conclusion. For example, a far higher percentage of visually impaired people have Labradors compared to the rest of the population, so you might jump to the conclusion that Labradors cause sight problems. Of course we know better, that if you are visually impaired then you will probably have a Labrador as a guide dog. To think otherwise is ridiculous.

But apply the same scenario to the complex human body and it is not always so transparent. Consequently, much of the debate about diet and nutrition is of the ‘chicken versus egg’ variety. Is a low or high amount of a nutrient a cause of a disease, a consequence of the disease, or simply irrelevant?

To try and limit this confusion, researchers often use what’s known as a cohort study. Say you’re interested in studying the effects of diet on cancer risk. You’d begin by taking a large population that are free of the disease at the outset and collect detailed data on their diet. You’d then follow this population over time, let’s say ten years, and see how many people were diagnosed with cancer during this period. You could then start to analyse the relationship between people’s diet and their risk of cancer, and ask a whole lot of interesting questions. Did people who ate a lot of fruit and veggies have less cancer? Did eating a lot of red meat increase cancer? What effect did drinking alcohol have on cancer risk? And so on.

The European Prospective Investigation into Cancer and Nutrition (EPIC), which we refer to often in this book, is an example of a powerfully designed cohort study, involving more than half a million people in ten countries. These studies are a gold mine of useful information because they help us piece together dietary factors that could influence our risk of disease.

But, however big and impressive these studies are, they’re still observational. As such they can only show us associations, they cannot prove causality. So if we’re not careful about the way we interpret this kind of research, we run the risk of drawing some whacky conclusions, just like we did with the Labradors. Let’s get back to some more news headlines, like this one we spotted: ‘Every hour per day watching TV increases risk of heart disease death by a fifth’.

When it comes to observational studies, you have to ask whether the association makes sense. Does it have ‘biological plausibility’? Are there harmful rays coming from the TV that damage our arteries or is it that the more time we spend on the couch watching TV, the less time we spend being active and improving our heart health. The latter is true, of course, and there’s an ‘association’ between TV watching and heart disease, not ‘causation’.

So even with cohorts, the champions of the epidemiological studies, we can’t prove causation, and that’s all down to what’s called ‘confounding’. This means there could be another variable at play that causes the disease being studied, at the same time as being associated with the risk factor being investigated. In our example, it’s the lack of physical activity that increases heart disease and is also linked to watching more TV.

This issue of confounding variables is just about the biggest banana skin of the lot. Time and time again you’ll find nutritional advice promoted on the basis of the findings of observational studies, as though this type of research gives us stone cold facts. It doesn’t. Any scientist will tell you that. This type of research is extremely useful for generating hypotheses, but it can’t prove them.

Rule #5: Be on the lookout for RCTs (randomized controlled trials)
An epidemiological study can only form a hypothesis, and when it offers up some encouraging findings, these then need to be tested in what’s known as an intervention, or clinical, trial before we can talk about causality. Intervention trials aim to test the hypothesis by taking a population that are as similar to each other as possible, testing an intervention on a proportion of them over a period of time and observing how it influences your measured outcome.