Chapter 11. Emotions, Aggression, and Stress
Follow us on Facebook and Twitter, or subscribe to our mailing list, to receive news updates. Learn more.
—By Chris Mooney The United States has a voting problem. In the 2012 presidential election, only about 57 percent of eligible American voters turned out, a far lower participation rate than in comparable democracies. That means about 93 million people who were eligible to vote didn't bother. Clearly, figuring out why people vote (and why they don't) is of premium importance to those who care about the health of democracy, as well as to campaigns that are becoming ever more sophisticated in targeting individual voters. To that end, much research has shown that demographic factors such as age and poverty affect one's likelihood of voting. But are there individual-level biological factors that also influence whether a person votes? The idea has long been heretical in political science, and yet the logic behind it is unavoidable. People vary in all sorts of ways—ranging from personalities to genetics—that affect their behavior. Political participation can be an emotional, and even a stressful activity, and in an era of GOP-led efforts to make voting more difficult, voting in certain locales can be a major hassle. To vote, you need both to be motivated and also not so intimidated you stay away from the polls. So are there biological factors that can shape these perceptions? "Our study is unique in that it is the first to examine whether differences in physiology may be causally related to differences in political activity," says lead study author Jeffrey French. ©2014 Mother Jones
Link ID: 19790 - Posted: 07.04.2014
From David Beckham’s infamous kick at France '98 to Luis Suárez chomping Giorgio Chiellini's shoulder in Brazil last week, the history of the World Cup is littered with moments of impulsive aggression that appear to defy all rational explanation. The story of human impulsivity stretches back deep into our evolutionary past. By nature, we are all prone to making quick, rash decisions that may lead to regret, and in some cases a four-month ban from international football. Impulsivity is actually a survival mechanism and was essential in the African savanna where our species evolved around a million and a half years ago. For our ancestors, the ability to make split-second decisions could make the difference between life and death. All of us have deep primal instincts but over the several hundred million years of evolution separating our reptilian ancestors from the first mammals, and eventually primates, the cognitive ability to exercise self-restraint has increased. While most living things make this decision purely as a trade-off between risk and reward, only humans can decide to exercise self-restraint on the basis of how they think they will be perceived by others – an ability that emerged some time in the past 100,000 years or so. “We evolved to be very social animals, living in large groups, and so we have developed inhibitory mechanisms in the more recently evolved parts of the prefrontal cortex,” explains Michael Price of the School of Social Sciences at the University of Brunel. “This is the social centre of the brain. Our big reason not to be impulsive is because of your reputation and how other people are going to judge you and perhaps ostracise you as we saw with Beckham in the aftermath of France ’98.” © 2014 Guardian News and Media Limited
By Tanya Lewis and Live Science They say laughter is the best medicine. But what if laughter is the disease? For a 6-year-old girl in Bolivia who suffered from uncontrollable and inappropriate bouts of giggles, laughter was a symptom of a serious brain problem. But doctors initially diagnosed the child with “misbehavior.” “She was considered spoiled, crazy — even devil-possessed,” José Liders Burgos Zuleta of the Advanced Medical Image Centre in La Paz said in a statement. [ But Burgos Zuleta discovered that the true cause of the girl’s laughing seizures, medically called gelastic seizures, was a brain tumor. After the girl underwent a brain scan, the doctors discovered a hamartoma, a small, benign tumor that was pressing against her brain’s temporal lobe. Surgeons removed the tumor, the doctors said. She stopped having the uncontrollable attacks of laughter and now laughs only normally, they said. Gelastic seizures are a relatively rare form of epilepsy, said Solomon Moshé, a pediatric neurologist at Albert Einstein College of Medicine in New York. “It’s not necessarily ‘ha-ha-ha’ laughing,” Moshé said. “There’s no happiness in this. Some of the kids may be very scared,” he added. The seizures are most often caused by tumors in the hypothalamus, although they can also come from tumors in other parts of brain, Moshé said. Although laughter is the main symptom, patients may also have outbursts of crying.
Sarah C. P. Williams There’s a reason people say “Calm down or you’re going to have a heart attack.” Chronic stress—such as that brought on by job, money, or relationship troubles—is suspected to increase the risk of a heart attack. Now, researchers studying harried medical residents and harassed rodents have offered an explanation for how, at a physiological level, long-term stress can endanger the cardiovascular system. It revolves around immune cells that circulate in the blood, they propose. The new finding is “surprising,” says physician and atherosclerosis researcher Alan Tall of Columbia University, who was not involved in the new study. “The idea has been out there that chronic psychosocial stress is associated with increased cardiovascular disease in humans, but what’s been lacking is a mechanism,” he notes. Epidemiological studies have shown that people who face many stressors—from those who survive natural disasters to those who work long hours—are more likely to develop atherosclerosis, the accumulation of fatty plaques inside blood vessels. In addition to fats and cholesterols, the plaques contain monocytes and neutrophils, immune cells that cause inflammation in the walls of blood vessels. And when the plaques break loose from the walls where they’re lodged, they can cause more extreme blockages elsewhere—leading to a stroke or heart attack. Studying the effect of stressful intensive care unit (ICU) shifts on medical residents, biologist Matthias Nahrendorf of Harvard Medical School in Boston recently found that blood samples taken when the doctors were most stressed out had the highest levels of neutrophils and monocytes. To probe whether these white blood cells, or leukocytes, are the missing link between stress and atherosclerosis, he and his colleagues turned to experiments on mice. © 2014 American Association for the Advancement of Science
Link ID: 19761 - Posted: 06.23.2014
by Laura Sanders Some brain cells need a jolt of stress to snap to attention. Cells called astroglia help regulate blood flow, provide energy to nearby cells and even influence messages’ movement between nerve cells. Now, scientists report June 18 in Neuron that astroglia can be roused by the stress molecule norepinephrine, an awakening that may help the entire brain jump into action. As mice were forced to walk on a treadmill, an activity that makes them alert, astroglia in several parts of their brains underwent changes in calcium levels, a sign of activity, neuroscientist Dwight Bergles of Johns Hopkins University School of Medicine and colleagues found. Norepinephrine, which acts as a fight-or-flight hormone in the body and a neural messenger in the brain, seemed to cause the cell activity boost. When researchers depleted norepinephrine, treadmill walking no longer activated astroglia. It’s not clear whether astroglia in all parts of the brain heed this wake-up call, nor is it clear whether this activation influences behavior. Norepinephrine might help shift brain cells, both neurons and astroglia, into a state of heightened vigilance, the authors write. © Society for Science & the Public 2000 - 2013.
By MARIA KONNIKOVA THE absurdity of having had to ask for an extension to write this article isn’t lost on me: It is, after all, a piece on time and poverty, or, rather, time poverty — about what happens when we find ourselves working against the clock to finish something. In the case of someone who isn’t otherwise poor, poverty of time is an unpleasant inconvenience. But for someone whose lack of time is just one of many pressing concerns, the effects compound quickly. We make a mistake when we look at poverty as simply a question of financial constraint. Take what happened with my request for an extension. It was granted, and the immediate time pressure was relieved. But even though I met the new deadline (barely), I’m still struggling to dig myself out from the rest of the work that accumulated in the meantime. New deadlines that are about to whoosh by, a growing list of ignored errands, a rent check and insurance payment that I just realized I haven’t mailed. And no sign of that promised light at the end of the tunnel. My experience is the time equivalent of a high-interest loan cycle, except instead of money, I borrow time. But this kind of borrowing comes with an interest rate of its own: By focusing on one immediate deadline, I neglect not only future deadlines but the mundane tasks of daily life that would normally take up next to no time or mental energy. It’s the same type of problem poor people encounter every day, multiple times: The demands of the moment override the demands of the future, making that future harder to reach. When we think of poverty, we tend to think about money in isolation: How much does she earn? Is that above or below the poverty line? But the financial part of the equation may not be the single most important factor. “The biggest mistake we make about scarcity,” Sendhil Mullainathan, an economist at Harvard who is a co-author of the book “Scarcity: Why Having Too Little Means So Much,” tells me, “is we view it as a physical phenomenon. It’s not.” © 2014 The New York Times Company
By JAMES GORMAN Crazed commuters, fretful parents and overwrought executives are not the only ones to suffer from anxiety — or to benefit from medication for it. The simple crayfish has officially entered the age of anxiety, too. This presumably was already clear to crayfish, which have been around for more than 200 million years and, what with predatory fish — and more recently, étouffée — have long had reasons to worry. But now scientists from France have documented behavior in crayfish that fits the pattern of a certain type of anxiety in human beings and other animals. Although the internal life of crayfish is still unknown, the findings, reported on Thursday in the journal Science, suggest that the external hallmarks of anxiety have been around for a very long time — and far down the food chain. Beyond that, a precursor of Valium changed the behavior back to normal. That does not mean that the crayfish are ready for the therapist’s couch, but it does reinforce the sometimes surprising connections humans have with other living things. Humans share genes with yeast as well as apes, the brains of flies can yield insights into the brains of humans, and even a tiny roundworm has mating behaviors that depend on a molecule very similar to a human hormone. The response to a threat or danger that the scientists found in crayfish had been documented before in other animals, like mice, but not in invertebrates like insects and crustaceans. Researchers including Pascal Fossat and Daniel Cattaert at the University of Bordeaux reported that after crayfish were exposed to electric shocks, they would not venture out of comfortable dark areas in an elaborate aquarium into scarier (for a crayfish) brightly lit areas. © 2014 The New York Times Company
The financial crisis has been linked to a 4.5 per cent increase in Canada’s suicide rate, according to a study that estimates at least 10,000 extra suicides could be connected to economic hardship in EU countries and North America. Researchers compared suicide data from the World Health Organization before and after the onset of the recession in 2007. "A crucial question for policy and psychiatric practice is whether these suicide rises are inevitable," Aaron Reeves of Oxford University’s sociology department and his co-authors said in Wednesday’s issue of the British Journal of Psychiatry. Given that the rise in suicides exceeded what would be expected and the large variations in suicide rates across countries, the researchers suspect some of the suicides were "potentially avoidable." In Canada, the suicides rose by 4.5 per cent or about 240 suicides more than expected between 2007 and 2010. In the U.S.A, the rate increased by 4.8 per cent over the same period. Before 2007 in Europe, suicide rates had been falling, but the trend reversed, rising by 6.5 per cent by 2009 and staying elevated through 2011. Two countries, Sweden and Finland, bucked the trend in the early 1990s. Job loss, home repossession and debt are the main risk factors leading to suicide during economic downturns, previous studies suggest. © CBC 2014
Jane J. Lee Could've, should've, would've. Everyone has made the wrong choice at some point in life and suffered regret because of it. Now a new study shows we're not alone in our reaction to incorrect decisions. Rats too can feel regret. Regret is thinking about what you should have done, says David Redish, a neuroscientist at the University of Minnesota in Minneapolis. It differs from disappointment, which you feel when you don't get what you expected. And it affects how you make decisions in the future. (See "Hand Washing Wipes Away Regrets?") If you really want to study emotions or feelings like regret, says Redish, you can't just ask people how they feel. So when psychologists and economists study regret, they look for behavioral and neural manifestations of it. Using rats is one way to get down into the feeling's neural mechanics. Redish and colleague Adam Steiner, also at the University of Minneapolis, found that rats expressed regret through both their behavior and their neural activity. Those signals, researchers report today in the journal Nature Neuroscience, were specific to situations the researchers set up to induce regret, which led to specific neural patterns in the brain and in behavior. When Redish and Steiner looked for neural activity, they focused on two areas known in people—and in some animals—to be involved in decision-making and the evaluation of expected outcomes: the orbitofrontal cortex and the ventral striatum. Brain scans have revealed that people with a damaged orbitofrontal cortex, for instance, don't express regret. To record nerve-cell activity, the researchers implanted electrodes in the brains of four rats—a typical sample size in this kind of experiment—then trained them to run a "choice" maze. © 1996-2014 National Geographic Society
By JOHN COATES SIX years after the financial meltdown there is once again talk about market bubbles. Are stocks succumbing to exuberance? Is real estate? We thought we had exorcised these demons. It is therefore with something close to despair that we ask: What is it about risk taking that so eludes our understanding, and our control? Part of the problem is that we tend to view financial risk taking as a purely intellectual activity. But this view is incomplete. Risk is more than an intellectual puzzle — it is a profoundly physical experience, and it involves your body. Risk by its very nature threatens to hurt you, so when confronted by it your body and brain, under the influence of the stress response, unite as a single functioning unit. This occurs in athletes and soldiers, and it occurs as well in traders and people investing from home. The state of your body predicts your appetite for financial risk just as it predicts an athlete’s performance. If we understand how a person’s body influences risk taking, we can learn how to better manage risk takers. We can also recognize that mistakes governments have made have contributed to excessive risk taking. Consider the most important risk manager of them all — the Federal Reserve. Over the past 20 years, the Fed has pioneered a new technique of influencing Wall Street. Where before the Fed shrouded its activities in secrecy, it now informs the street in as clear terms as possible of what it intends to do with short-term interest rates, and when. Janet L. Yellen, the chairwoman of the Fed, declared this new transparency, called forward guidance, a revolution; Ben S. Bernanke, her predecessor, claimed it reduced uncertainty and calmed the markets. But does it really calm the markets? Or has eliminating uncertainty in policy spread complacency among the financial community and actually helped inflate market bubbles? We get a fascinating answer to these questions if we turn from economics and look into the biology of risk taking. © 2014 The New York Times Company
By Jonathan Webb Science reporter, BBC News A new theory suggests that our male ancestors evolved beefy facial features as a defence against fist fights. The bones most commonly broken in human punch-ups also gained the most strength in early "hominin" evolution. They are also the bones that show most divergence between males and females. The paper, in the journal Biological Reviews, argues that the reinforcements evolved amid fighting over females and resources, suggesting that violence drove key evolutionary changes. For many years, this extra strength was seen as an adaptation to a tough diet including nuts, seeds and grasses. But more recent findings, examining the wear pattern and carbon isotopes in australopith teeth, have cast some doubt on this "feeding hypothesis". "In fact, [the australopith] boisei, the 'nutcracker man', was probably eating fruit," said Prof David Carrier, the new theory's lead author and an evolutionary biologist at the University of Utah. Masculine armour Instead of diet, Prof Carrier and his co-author, physician Dr Michael Morgan, propose that violent competition demanded the development of these facial fortifications: what they call the "protective buttressing hypothesis". In support of their proposal, Carrier and Morgan offer data from modern humans fighting. Several studies from hospital emergency wards, including one from the Bristol Royal Infirmary, show that faces are particularly vulnerable to violent injuries. BBC © 2014
Neil Levy Can human beings still be held responsible in the age of neuroscience? Some people say no: they say once we understand how the brain processes information and thereby causes behaviour, there’s nothing left over for the person to do. This argument has not impressed philosophers, who say there doesn’t need to be anything left for the person to do in order to be responsible. People are not anything over and above the causal systems involved in information processing, we are our brains (plus some other, equally physical stuff). We are responsible if our information processing systems are suitably attuned to reasons, most philosophers think. There are big philosophical debates concerning what it takes to be suitably attuned to reasons, and whether this is really enough for responsibility. But I want to set those debates aside here. It’s more interesting to ask what we can learn from neuroscience about the nature of responsibility and about when we’re responsible. Even if neuroscience doesn’t tell us that no one is ever responsible, it might be able to tell us if particular people are responsible for particular actions. A worthy case study Consider a case like this: early one morning in 1987, a Canadian man named Ken Parks got up from the sofa where he had fallen asleep and drove to his parents’-in-law house. There he stabbed them both before driving to the police station, where he told police he thought he had killed someone. He had: his mother-in-law died from her injuries. © 2010–2014, The Conversation Trust (UK)
By Denali Tietjen Meditation has long been known for its mental health benefits, but new research shows that just a few minutes of mindfulness can improve physical health and personal life as well. A recent study conducted by researchers at INSEAD and The Wharton School found that 15 minutes of mindful meditation can help you make better decisions. The research, published in the Association for Psychological Science’s journal Psychological Science, comes from four studies (varying in sample size from 69 to 178 adults) in which participants responded to sunk-cost scenarios at different degrees of mindful awareness. The results consistently showed that increased mindfulness decreases the sunk-cost bias. WOAH, hold the phone. What’s a sunk cost and what’s a sunk-cost bias?? Sunk cost is an economics term that psychologists have adopted. In economics, sunk costs are defined as non-recoverable investment costs like the cost of employee training or a lease on office space. In psychology, sunk costs are basically the same thing: The time and energy we put into our personal lives. Though we might not sit down with a calculator at the kitchen table when deciding who to take as our plus one to our second cousin’s wedding next weekend, we do a cost-benefit analysis every time we make a decision. And we take these sunk costs into account. The sunk-cost bias, then, is the tendency to allow sunk costs to overly influence current decisions. Mindfulness meditation can provide improved clarity, which helps you stay present and make better decisions, the study says. This protects you from that manipulative sunk-cost bias.
Link ID: 19693 - Posted: 06.05.2014
by Bethany Brookshire We all respond to stress in different ways. Some of us work harder. Others drink more or eat our feelings. Sometimes we experience sleep loss, heart palpitations or sweats. When the stress dissipates, many of us go back to our daily lives, none the worse for wear. We are resilient. But some people find that stress is a first step on the way to a major depressive episode. It’s not quite clear what’s different between people who go back to normal after stress, and those who descend into depression. “One of the most important questions is, how do the brains of resilient animals (or humans) differ from those that are vulnerable to depression following stress?” asks John Morrison, a neuroscientist at the Icahn School of Medicine at Mount Sinai in New York. A new study from Minghui Wang and colleagues at Cold Spring Harbor Laboratory in New York provides a new hint. Mice with a depressive-like response to stress have stronger connections between neurons in the medial prefrontal cortex of the brain following the stress. Resilient mice show weaker connections. The mechanism could help scientists understand why some people respond to stress with depression, while others are able to shake it off. The prefrontal cortex is best known for its role in executive function — thought, memory, prediction and other tasks. But dysfunction in some areas of the cortex, particularly one called Brodmann area 25, has been linked with recurring major depressive disorder. Scientists have been electrically stimulating this area to relieve depression in patients. But researchers still don’t understand what makes this brain area important in depression, and how dysfunctions might occur. “I’ve had a long interest in the mechanism of human diseases like depression,” says study coauthor Bo Li, a cellular and behavioral neuroscientist at Cold Spring Harbor. “The idea has been to identify an area that is responsible, to link a mechanism in the brain to a behavior.” Wang, Li and their colleagues were especially interested in changes to the mouse prefrontal cortex following stress. © Society for Science & the Public 2000 - 2013.
By GRETCHEN REYNOLDS If you are aiming to lose weight by revving up your exercise routine, it may be wise to think of your workouts not as exercise, but as playtime. An unconventional new study suggests that people’s attitudes toward physical activity can influence what they eat afterward and, ultimately, whether they drop pounds. For some time, scientists have been puzzled — and exercisers frustrated — by the general ineffectiveness of exercise as a weight-loss strategy. According to multiple studies and anecdotes, most people who start exercising do not lose as much weight as would be expected, given their increased energy expenditure. Some people add pounds despite burning hundreds of calories during workouts. Past studies of this phenomenon have found that exercise can increase the body’s production of appetite hormones, making some people feel ravenous after even a light workout and prone to consume more calories than they expended. But that finding, while intriguing, doesn’t fully explain the wide variability in people’s post-exercise eating habits. So, for the new study, published in the journal Marketing Letters, French and American researchers turned to psychology and the possible effect that calling exercise by any other name might have on people’s subsequent diets. In that pursuit, the researchers first recruited 56 healthy, adult women, the majority of them overweight. The women were given maps detailing the same one-mile outdoor course and told that they would spend the next half-hour walking there, with lunch to follow. Half of the women were told that their walk was meant to be exercise, and they were encouraged to view it as such, monitoring their exertion throughout. The other women were told that their 30-minute outing would be a walk purely for pleasure; they would be listening to music through headphones and rating the sound quality, but mostly the researchers wanted them to enjoy themselves. When the women returned from walking, the researchers asked each to estimate her mileage, mood and calorie expenditure. © 2014 The New York Times Company
Ewen Callaway By controlling rats' brain cells they had genetically engineered to respond to light, researchers were able to create fearful memories of events that never happened — and then to erase those memories again. Neuroscientists can breathe a collective sigh of relief. Experiments have confirmed a long-standing theory for how memories are made and stored in the brain. Researchers have created and erased frightening associations in rats' brains using light, providing the most direct demonstration yet that the strengthening and weakening of connections between neurons is the basis for memory. “This is the best evidence so far available, period,” says Eric Kandel, a neuroscientist at Columbia University in New York. Kandel, who shared the 2000 Nobel Prize in Physiology or Medicine for his work unravelling the molecular basis of memory, was not involved in the latest study, which was published online in Nature1 on 1 June. In the 1960s and 1970s, researchers in Norway noticed a peculiar property of brain cells. Repeatedly delivering a burst of electricity to a neuron in an area of the brain known as the hippocampus seemed to boost the cell’s ability to talk to a neighbouring neuron. These communiqués occur across tiny gaps called synapses, which neurons can form with thousands of other nerve cells. The process was called long-term potentiation (LTP), and neuroscientists suspected that it was the physical basis of memory. The hippocampus, they realized, was important for forming long-term memories, and the long-lasting nature of LTP hinted that information might be stored in a neural circuit for later recall. © 2014 Nature Publishing Group,
By Lara Salahi A team of researchers at Massachusetts General Hospital have embarked on a new project to create an implantable device in the brain that would read and alter the emotions of someone with a mental illness. The team is working in collaboration with researchers at the University of California, San Francisco, on a new program funded by the Department of Defense’s Defense Advanced Research Projects Agency (DARPA). The researchers are working to create an implantable device that can sense abnormal activity in the brain using algorithms, and then deliver electrical impulses to certain parts of the brain that would suppress the abnormal signals. “Imagine if I have an addiction to alcohol and I have a craving,” Jose Carmena, a researcher at the University of California, Berkeley, who is involved in the project, told MIT Technology Review. “We could detect that feeling and then stimulate inside the brain to stop it from happening.” Mental illness and suicide rates among the US military have spiked over the past decade, the National Institute of Mental Healthreports. The current research is part of DARPA’s emerging neurotechnology therapy program which investigates new approaches to treat neuropsychological illnesses among military servicemembers and veterans. Their goal is to treat at least seven psychiatric conditions, including depression, post-traumatic stress disorder, addiction, and fibromyalgia.
Jyoti Madhusoodanan Most people handle stress well, but some find it difficult to cope and as a result develop depression and other mood disorders. Researchers have previously been able to identify the part of the brain that controls this response, but not exactly how it does so. Now, a study in mice identifies a small group of neurons that could be responsible. The research might also help elucidate the mechanism of deep brain stimulation, a therapy that uses electrical impulses to treat depression and other neurological disorders. How an animal deals with stress is controlled by a part of the brain known as the prefrontal cortex, and the neurons in this part of the brain are known to change in structure and function in response to stressful situations1. To look at the cellular basis of the responses, neuroscientist Bo Li of Cold Spring Harbor Laboratory in New York and his colleagues subjected mice to small electric shocks at random intervals to produce stress. Most of the mice tried to avoid the shocks, but just over one-fifth did not. They also started to avoid other animals or failed to choose tasty foods over plain ones — typical signs of depressive behaviour. The researchers then looked at the animals' brains and found that a specific set of neurons in the prefrontal cortex were easily excitable in depressed mice, but much harder to excite in those resilient to the stress. Furthermore, artificially increasing the activity of these neurons caused mice that were once resilient to become susceptible to depressive behaviours. “We were surprised that we were able to see a difference between depressed and resilient animals at the level of synaptic transmission,” says Li. © 2014 Nature Publishing Group,
|By Bret Stetka Skepticism around fibromyalgia stemmed in part from an elusive organic explanation. Symptoms appeared to arise out of nowhere, which didn't make any sense to empirically minded physicians. “I, too, have been assigned months of futility, long and weary nights of misery. When I go to bed, I think,`When will it be morning?' But the night drags on, and I toss till dawn...Depression haunts my days. My weary nights are filled with pain as though something were relentlessly gnawing at my bones.” Job suffered badly. And his Old Testament woes are considered by many to be one of the earliest descriptions of fibromyalgia, a painful, puzzling disorder that still has experts bickering and patients frustrated, bereft of relief. The Bible isn't exactly a paragon of medical accuracy, but Job’s ailment does sound an awful lot like the modern interpretation of fibromyalgia. The classic diffuse pain, aches and discomfort aren’t the half of it; depression, fatigue, stiffness, sleep loss and generally just feeling really bad are common too. Fibromyalgia patients — 2 percent to 8 percent of the population — have also endured decades of dismissals that it's all in their head — a psychosomatic conjuring, a failure of constitution. Skepticism around fibromyalgia stemmed in part from an elusive organic explanation. Symptoms appeared to arise out of nowhere, which didn't make any sense to empirically minded physicians. But over the past two decades, research has brought clinicians closer to deciphering this mysterious pain state, once thought muscular in nature, now known to be neurologic. Based on this recent work a new article in the Journal of the American Medical Association by chronic pain expert Dr. Daniel Clauw brings us up to speed on the understanding, diagnosis and management of fibromyalgia circa 2014. And the outlook for patients is rosier than you might expect given the condition’s perplexing reputation. © 2014 Scientific American
Claudia M. Gold When Frank was a young boy, and he committed some typical toddler transgression such as having a meltdown when it was time to leave the playground, his father would slap him across the face, hurting and humiliating him in a very public way. When I spoke with Frank over 20 years later, in the context of helping him with his own son Leo's frequent tantrums in my behavioral pediatrics practice, he did not describe this experience as "trauma." Rather, he described it in a very matter-of-fact tone. But when we explored in detail his response to his son's tantrums, we discovered that, flooded by the stress of his own memories, Frank in a sense would shut down. Normally a thoughtful and empathic person, he simply told Leo to "cut it out." As we spoke he recognized how he was emotionally absent during these moments, which were increasing in frequency. It seemed as if Leo was testing Frank, perhaps looking for a more appropriate response that would help him manage this normal behavior. Once this process was brought in to awareness, Frank was able to be present with Leo- to tolerate his tantrums and understand them from his 2-year-old perspective. Soon the frequency and intensity of the tantrums returned to a level typical for Leo's developmental stage. Frank, greatly relieved, once again found himself enjoying his son. The upcoming Boston conference; Psychological Trauma: Neuroscience, Attachment, and Therapeutic Interventions, promises to offer insight in to the developmental neuroscience behind this story. What Frank experienced as a young child might be termed "quotidian" or "everyday" trauma. It was not watching a relative get shot, or having his house washed away in an avalanche. It was a daily mismatch with his father- he was looking for reassurance and containment and instead got a slap across the face. It was what leading researcher Ed Tronick would term "unrepaired mismatch." Frank, in a way that is extremely common- termed "intergenerational transmission of trauma"- was then repeating this cycle with his own child. When this dynamic was brought in to awareness, he was able to "repair the mismatch," setting his relationship with his own son on a healthier path. ©2014 Boston Globe Media Partners, LLC