Chapter 14. Attention and Consciousness
Follow us on Facebook and Twitter, or subscribe to our mailing list, to receive news updates. Learn more.
by Katie Collins Sarah-Jayne Blakemore is just as fascinated by the links between neuroscience and education as she is outraged by the pseudo science that often intrudes upon this territory. Neuroscience in education has really been flourishing in recent years, she says on stage at WIRED Health 2015, but some theories about neuroscience have already infiltrated schools, and not necessarily in a good way. Some products that makes claims about having a positive effect on cognition make bogus claims that may well have positive effects in the classroom, but at the same time promote completely inaccurate science. Blakemore points specifically to the Brain Gym educational model, which claims to improve memory, concentration and information retention. There are no problems with the exercises themselves, she says, but the claims made about the brain are baseless. For a start, she said, Brain Gym claims that children can push "brain buttons" on their bodies that will stimulate blood flow to the brain. Another physical exercise claimed to increase and improve connectivity between the two sides of the brain. "This makes no sense -- they are in communication anyway," says Blakemore. Teachers like Brain Gym because it does what it says and results in improvements in the classroom, but it could just as easily be placebo or novelty causing the effects. One thing Blakemore is sure of? "They're nothing to do with brain buttons or coordinating the two brain hemispheres."
Keyword: Development of the Brain
Link ID: 20844 - Posted: 04.25.2015
By Jerry Adler Smithsonian Magazine | In London, Benjamin Franklin once opened a bottle of fortified wine from Virginia and poured out, along with the refreshment, three drowned flies, two of which revived after a few hours and flew away. Ever the visionary, he wondered about the possibility of incarcerating himself in a wine barrel for future resurrection, “to see and observe the state of America a hundred years hence.” Alas, he wrote to a friend in 1773, “we live in an age too early . . . to see such an art brought in our time to its perfection.” If Franklin were alive today he would find a kindred spirit in Ken Hayworth, a neuroscientist who also wants to be around in 100 years but recognizes that, at 43, he’s not likely to make it on his own. Nor does he expect to get there preserved in alcohol or a freezer; despite the claims made by advocates of cryonics, he says, the ability to revivify a frozen body “isn’t really on the horizon.” So Hayworth is hoping for what he considers the next best thing. He wishes to upload his mind—his memories, skills and personality—to a computer that can be programmed to emulate the processes of his brain, making him, or a simulacrum, effectively immortal (as long as someone keeps the power on). Hayworth’s dream, which he is pursuing as president of the Brain Preservation Foundation, is one version of the “technological singularity.” It envisions a future of “substrate-independent minds,” in which human and machine consciousness will merge, transcending biological limits of time, space and memory. “This new substrate won’t be dependent on an oxygen atmosphere,” says Randal Koene, who works on the same problem at his organization, Carboncopies.org. “It can go on a journey of 1,000 years, it can process more information at a higher speed, it can see in the X-ray spectrum if we build it that way.”
|By Tara Haelle When it comes to treating attention-deficit hyperactivity disorder (ADHD) a lot of kids are getting the meds they need—but they may be missing out on other treatments. Despite clinical guidelines that urge that behavioral therapy always be used alongside medication, less than half of the children with ADHD received therapy as part of treatment in 2009 and 2010, according to the first nationally representative study of ADHD treatment in U.S. children. The findings, published online March 31 in The Journal of Pediatrics, come from data collected during that period on 9,459 children, aged four to 17, with diagnosed ADHD—just before the American Academy of Pediatrics (AAP) issued its clinical practice guidelines on treatments of the condition in 2011. They provide a baseline for comparison when the next report is issued in 2017. Medication alone was the most common treatment for children with ADHD: 74 percent had taken medication in the previous week whereas 44 percent had received behavioral therapy in the past year. Just under a third of children of all ages had received both medication and behavioral therapy, the AAP-recommended treatment for all ages. “It’s not at all surprising that medication is the most common treatment,” says Heidi Feldman, a professor of developmental and behavioral pediatrics at Stanford University School of Medicine who served on the AAP clinical practice guidelines committee. “It works very effectively to reduce the core symptoms of the condition,” she adds, “and stimulants are relatively safe if used properly. The limitation of stimulant medications for ADHD is that studies do not show a long-term functional benefit from medication use.” © 2015 Scientific American
Link ID: 20827 - Posted: 04.21.2015
By ALAN SCHWARZ Fading fast at 11 p.m., Elizabeth texted her dealer and waited just 30 minutes for him to reach her third-floor New York apartment. She handed him a wad of twenties and fifties, received a tattered envelope of pills, and returned to her computer. Her PowerPoint needed another four hours. Investors in her health-technology start-up wanted re-crunched numbers, a presentation begged for bullet points and emails from global developers would keep arriving well past midnight. She gulped down one pill — pale orange, like baby aspirin — and then, reconsidering, took one of the pinks, too. “O.K., now I can work,” Elizabeth exhaled. Several minutes later, she felt her brain snap to attention. She pushed her glasses up her nose and churned until 7 a.m. Only then did she sleep for 90 minutes, before arriving at her office at 9. The pills were versions of the drug Adderall, an amphetamine-based stimulant prescribed for attention deficit hyperactivity disorder that many college students have long used illicitly while studying. Now, experts say, stimulant abuse is graduating into the work force. But in interviews, dozens of people in a wide spectrum of professions said they and co-workers misused stimulants like Adderall, Vyvanse and Concerta to improve work performance. Most spoke on the condition of anonymity for fear of losing their jobs or access to the medication. Doctors and medical ethicists expressed concern for misusers’ health, as stimulants can cause anxiety, addiction and hallucinations when taken in high doses. But they also worried about added pressure in the workplace — where the use by some pressures more to join the trend. © 2015 The New York Times Company
by Anil Ananthaswamy HOLD that thought. When it comes to consciousness, the brain may be doing just that. It now seems that conscious perception requires brain activity to hold steady for hundreds of milliseconds. This signature in the pattern of brainwaves can be used to distinguish between levels of impaired consciousness in people with brain injury. The new study by Aaron Schurger at the Swiss Federal Institute of Technology in Lausanne doesn't explain the so-called "hard problem of consciousness" – how roughly a kilogram of nerve cells is responsible for the miasma of sensations, thoughts and emotions that make up our mental experience. However, it does chip away at it, and support the idea that it may one day be explained in terms of how the brain processes information. Neuroscientists think that consciousness requires neurons to fire in such a way that they produce a stable pattern of brain activity. The exact pattern will depend on what the sensory information is, but once information has been processed, the idea is that the brain should hold a pattern steady for a short period of time – almost as if it needs a moment to read out the information. In 2009, Schurger tested this theory by scanning 12 people's brains with fMRI machines. The volunteers were shown two images simultaneously, one for each eye. One eye saw a red-on-green line drawing and the other eye saw green-on-red. This confusion caused the volunteers to sometimes consciously perceive the drawing and sometimes not. © Copyright Reed Business Information Ltd.
By Shereen Lehman (Reuters Health) - Children exposed to tobacco smoke at home are up to three times more likely to have attention deficit hyperactive disorder (ADHD) as unexposed kids, according to a new study from Spain. The association was stronger for kids with one or more hours of secondhand smoke exposure every day, the authors found. And the results held when researchers accounted for parents' mental health and other factors. "We showed a significant and substantial dose-response association between (secondhand smoke) exposure in the home and a higher frequency of global mental problems," the authors write in Tobacco Control, online March 25. According to the Centers for Disease Control and Prevention, two of every five children in the US are exposed to secondhand smoke regularly. Alicia Padron of the University of Miami Miller School of Medicine in Florida and colleagues in Spain analyzed data from the 2011 to 2012 Spanish National Health Interview Survey, in which parents of 2,357 children ages four to 12 reported the amount of time their children were exposed to secondhand smoke every day. The parents also filled out questionnaires designed to evaluate their children's mental health. According to the results, about eight percent of the kids had a probable mental disorder. About 7% of the kids were exposed to secondhand smoke for less than one hour per day, and 4.5% were exposed for an hour or more each day. © 2015 Scientific American,
By Lawrence Berger A cognitive scientist and a German philosopher walk into the woods and come upon a tree in bloom: What does each one see? And why does it matter? While that may sound like the set-up to a joke making the rounds at a philosophy conference, I pose it here sincerely, as a way to explore the implications of two distinct strains of thought — that of cognitive science and that of phenomenology, in particular, the thought of Martin Heidegger, who offers a most compelling vision of the ultimate significance of our being here, and what it means to be fully human. When we feel that someone is really listening to us, we feel more alive, we feel our true selves coming to the surface — this is the sense in which worldly presence matters. It can be argued that cognitive scientists tend to ignore the importance of what many consider to be essential features of human existence, preferring to see us as information processors rather than full-blooded human beings immersed in worlds of significance. In general, their intent is to explain human activity and life as we experience it on the basis of physical and physiological processes, the implicit assumption being that this is the domain of what is ultimately real. Since virtually everything that matters to us as human beings can be traced back to life as it is experienced, such thinking is bound to be unsettling. For instance, an article in The Times last year by Michael S. A. Graziano, a professor of psychology and neuroscience at Princeton, about whether we humans are “really conscious,” argued, among other things, that “we don’t actually have inner feelings in the way most of us think we do.” © 2015 The New York Times Company
by Michael Slezak What were we talking about? Oh yes, brain-training programmes may be useful for helping inattentive people focus on tasks in their daily life. At least, that's the implication of an analysis looking at one particular programme. It's the latest salvo in a field that has seen the battles lines drawn between those who believe there is no compelling scientific evidence that training the brain to do a specific task better can offer wider cognitive improvements, and those that think it can work in some cases. The party line is that brain training improves only that which it exercises, says Jared Horvath from the University of Melbourne in Australia. "What this means is, if the training programme uses a working memory game, you get better at working memory games and little else." But an analysis by Megan Spencer-Smith of Monash University in Melbourne, Australia, and Torkel Klingberg at the Karolinska Institute in Stockholm, Sweden, claims to show that there are benefits for daily life – at least for people with attention deficit hyperactivity disorder or other problems related to attentiveness. They focused on a programme called Cogmed, which Klingberg has helped develop, and combined the results of several smaller studies. Cogmed is designed to improve how much verbal or visual information you can temporarily remember and work with. © Copyright Reed Business Information Ltd.
By Nicholas Weiler Where did the thief go? You might get a more accurate answer if you ask the question in German. How did she get away? Now you might want to switch to English. Speakers of the two languages put different emphasis on actions and their consequences, influencing the way they think about the world, according to a new study. The work also finds that bilinguals may get the best of both worldviews, as their thinking can be more flexible. Cognitive scientists have debated whether your native language shapes how you think since the 1940s. The idea has seen a revival in recent decades, as a growing number of studies suggested that language can prompt speakers to pay attention to certain features of the world. Russian speakers are faster to distinguish shades of blue than English speakers, for example. And Japanese speakers tend to group objects by material rather than shape, whereas Koreans focus on how tightly objects fit together. Still, skeptics argue that such results are laboratory artifacts, or at best reflect cultural differences between speakers that are unrelated to language. In the new study, researchers turned to people who speak multiple languages. By studying bilinguals, “we’re taking that classic debate and turning it on its head,” says psycholinguist Panos Athanasopoulos of Lancaster University in the United Kingdom. Rather than ask whether speakers of different languages have different minds, he says, “we ask, ‘Can two different minds exist within one person?’ ” Athanasopoulos and colleagues were interested in a particular difference in how English and German speakers treat events. © 2015 American Association for the Advancement of Science
Brian Owens Our choice between two moral options might be swayed by tracking our gaze, and asking for a decision at the right moment. People asked to choose between two written moral statements tend to glance more often towards the option they favour, experimental psychologists say. More surprisingly, the scientists also claim it’s possible to influence a moral choice: asking for an immediate decision as soon as someone happens to gaze at one statement primes them to choose that option. It’s well known that people tend to look more towards the option they are going to choose when they are choosing food from a menu, says Philip Pärnamets, a cognitive scientist from Lund University in Sweden. He wanted to see if that applied to moral reasoning as well. “Moral decisions have long been considered separately from general decision-making,” he says. “I wanted to integrate them.” In a paper published today in the Proceedings of the National Academy of Sciences1, Pärnamets and his colleagues explain how they presented volunteers with a series of moral statements, such as 'murder is sometimes justified,' 'masturbating with the aid of a willing animal is acceptable' and 'paying taxes is a good thing.' Then the psychologists tracked the volunteers’ gaze as two options appeared on a screen. Once the tracker had determined that a person had spent at least 750 milliseconds looking at one answer and 250 milliseconds at the other, the screen changed to prompt them to make a decision. Almost 60% of the time, they chose the most viewed option — indicating, says Pärnamets, that eye gaze tracks an unfolding moral decision. © 2015 Nature Publishing Group,
|By Esther Landhuis As we age, we seem to get worse at ignoring irrelevant stimuli. It's what makes restaurant conversations challenging—having to converse while also shutting out surrounding chatter. New research bears out the aging brain's distractibility but also suggests that training may help us tune out interference. Scientists at Brown University recruited seniors and twentysomethings for a visual experiment. Presented with a sequence of letters and numbers, participants were asked to report back only the numbers—all the while disregarding a series of meaningless dots. Sometimes the dots moved randomly, but other times they traveled in a clear direction, making them harder to ignore. Older participants ended up accidentally learning the dots' patterns, based on the accuracy of their answers when asked which way the dots were moving, whereas young adults seemed able to suppress that information and focus on the numbers, the researchers reported last November in Current Biology. In a separate study published in Neuron, scientists at the University of California, San Francisco, showed they could train aging brains to become less distractible. Their regimen helped aging rats as well as older people. The researchers played three different sounds and rewarded trainees for identifying a target tone while ignoring distracter frequencies. As the subjects improved, the task grew more challenging—the distracting tone became harder to discriminate from the target. © 2015 Scientific American,
Robin Tricoles The first time it happened, I was 8. I was tucked in bed reading my favorite book when my tongue swelled up to the size of a cow’s, like the giant tongues I had seen in the glass display case at the neighborhood deli. At the same time, the far wall of my bedroom began to recede, becoming a tiny white rectangle floating somewhere in the distance. In the book I was holding, the typeface grew vast on the page. I was intrigued, I remember, but not afraid. Over the next six years, the same thing happened to me dozens of times. Forty years later, while working as a science writer, I stumbled on a scientific paper describing almost exactly what I had experienced. The paper attributed those otherworldly sensations to something called Alice in Wonderland syndrome, or its close cousin, Alice in Wonderland-like syndrome. People with Alice in Wonderland syndrome (AWS) perceive parts of their body to be changing size. For example, their feet may suddenly appear smaller and more distant, or their hands larger than they had been moments before. Those with the closely related Alice in Wonderland-like syndrome (AWLS) misperceive the size and distance of objects, seeing them as startlingly larger, smaller, fatter, or thinner than their natural state. People who experience both sensations, like I did, are classified as having AWLS. The syndrome’s name is commonly attributed to English psychiatrist John Todd, who in 1955 described his adult patients’ illusions of corporal and objective distortions in a paper in the Canadian Medical Association Journal. © 2015 by The Atlantic Monthly Group.
Link ID: 20672 - Posted: 03.10.2015
By TIMOTHY WILLIAMS In January 1972, Cecil Clayton was cutting wood at his family’s sawmill in southeastern Missouri when a piece of lumber flew off the circular saw blade and struck him in the forehead. The impact caved in part of Mr. Clayton’s skull, driving bone fragments into his brain. Doctors saved his life, but in doing so had to remove 20 percent of his frontal lobe, which psychiatrists say led Mr. Clayton to be tormented for years by violent impulses, schizophrenia and extreme paranoia. In 1996, his lawyers say, those impulses drove Mr. Clayton to kill a law enforcement officer. Today, as Mr. Clayton, 74, sits on death row, his lawyers have returned to that 1972 sawmill accident in a last-ditch effort to save his life, arguing that Missouri’s death penalty law prohibits the execution of severely brain-damaged people. Lawyers for Mr. Clayton, who has an I.Q. of 71, say he should be spared because his injury has made it impossible for him to grasp the significance of his death sentence, scheduled for March 17. “There was a profound change in him that he doesn’t understand, and neither did his family,” said Elizabeth Unger Carlyle, one of Mr. Clayton’s lawyers. While several rulings by the United States Supreme Court in recent years have narrowed the criteria for executing people who have a mental illness, states continue to hold wide sway in establishing who is mentally ill. The debate surrounding Mr. Clayton involves just how profoundly his impairment has affected his ability to understand what is happening to him. Mr. Clayton is missing about 7.7 percent of his brain. © 2015 The New York Times Company
|By Christof Koch In the Dutch countryside, a tall, older man, dressed in a maroon sports coat, his back slightly stooped, stands out because of his height and a pair of extraordinarily bushy eyebrows. His words, inflected by a British accent, are directed at a middle-aged man with long, curly brown hair, penetrating eyes and a dark, scholarly gown, who talks in only a halting English that reveals his native French origins. Their strangely clashing styles of speaking and mismatched clothes do not seem to matter to them as they press forward, with Eyebrows peering down intently at the Scholar. There is something distinctly odd about the entire meeting—a crossing of time, place and disciplines. Eyebrows: So I finally meet the man who doubts everything. The Scholar: (not missing a beat) At this time, I admit nothing that is not necessarily true. I'm famous for that! Eyebrows: Is there anything that you are certain of? (sotto voce) Besides your own fame? The Scholar: (evading the sarcastic jibe) I can't be certain of my fame. Indeed, I can't even be certain that there is a world out there, for I could be dreaming or hallucinating it. I can't be certain about the existence of my own body, its shape and extension, its corporality, for again I might be fooling myself. But now what am I, when I suppose that there is some supremely powerful and, if I may be permitted to say so, malicious deceiver who deliberately tries to fool me in any way he can? Given this evil spirit, how do I know that my sensations about the outside world—that is, it looks, feels and smells in a particular way—are not illusions, conjured up by Him to deceive me? It seems to me that therefore I can never know anything truly about the world. Nothing, rien du tout. I have to doubt everything. © 2015 Scientific American
Link ID: 20640 - Posted: 03.03.2015
By Neuroskeptic In an interesting short paper just published in Trends in Cognitive Science, Caltech neuroscientist Ralph Adolphs offers his thoughts on The Unsolved Problems of Neuroscience. Here’s Adolphs’ list of the top 23 questions (including 3 “meta” issues), which, he says, was inspired by Hilbert’s famous set of 23 mathematical problems: Problems that are solved, or soon will be: I. How do single neurons compute? II. What is the connectome of a small nervous system, like that of Caenorhabitis elegans (300 neurons)? III. How can we image a live brain of 100,000 neurons at cellular and millisecond resolution? IV. How does sensory transduction work? Problems that we should be able to solve in the next 50 years: V. How do circuits of neurons compute? VI. What is the complete connectome of the mouse brain (70,000,000 neurons)? VII. How can we image a live mouse brain at cellular and millisecond resolution? VIII. What causes psychiatric and neurological illness? IX. How do learning and memory work? X. Why do we sleep and dream? XI. How do we make decisions? XII. How does the brain represent abstract ideas? Problems that we should be able to solve, but who knows when: XIII. How does the mouse brain compute? XIV. What is the complete connectome of the human brain (80,000,000,000 neurons)? XV. How can we image a live human brain at cellular and millisecond resolution? XVI. How could we cure psychiatric and neurological diseases? XVII. How could we make everybody’s brain function best? Problems we may never solve: XVIII. How does the human brain compute? XIX. How can cognition be so flexible and generative? XX. How and why does conscious experience arise? Meta-questions: XXI. What counts as an explanation of how the brain works? (and which disciplines would be needed to provide it?) XXII. How could we build a brain? (how do evolution and development do it?) XXIII. What are the different ways of understanding the brain? (what is function, algorithm, implementation?) Adolphs R (2015). The unsolved problems of neuroscience. Trends in cognitive sciences PMID: 25703689
Link ID: 20637 - Posted: 03.02.2015
By Adam Rogers The fact that a single image could polarize the entire Internet into two aggressive camps is, let’s face it, just another Thursday. But for the past half-day, people across social media have been arguing about whether a picture depicts a perfectly nice bodycon dress as blue with black lace fringe or white with gold lace fringe. And neither side will budge. This fight is about more than just social media—it’s about primal biology and the way human eyes and brains have evolved to see color in a sunlit world. Light enters the eye through the lens—different wavelengths corresponding to different colors. The light hits the retina in the back of the eye where pigments fire up neural connections to the visual cortex, the part of the brain that processes those signals into an image. Critically, though, that first burst of light is made of whatever wavelengths are illuminating the world, reflecting off whatever you’re looking at. Without you having to worry about it, your brain figures out what color light is bouncing off the thing your eyes are looking at, and essentially subtracts that color from the “real” color of the object. “Our visual system is supposed to throw away information about the illuminant and extract information about the actual reflectance,” says Jay Neitz, a neuroscientist at the University of Washington. “But I’ve studied individual differences in color vision for 30 years, and this is one of the biggest individual differences I’ve ever seen.” (Neitz sees white-and-gold.) Usually that system works just fine. This image, though, hits some kind of perceptual boundary. That might be because of how people are wired. Human beings evolved to see in daylight, but daylight changes color. WIRED.com © 2015 Condé Nast
by Helen Thomson We meet in a pub, we have a few drinks, some dinner and then you lean in for a kiss. You predict, based on our previous interactions, that the kiss will be reciprocated – rather than landing you with a slap in the face. All our social interactions require us to anticipate another person's undecided intentions and actions. Now, researchers have discovered specific brain cells that allow monkeys to do this. It is likely that the cells do the same job in humans. Keren Haroush and Ziv Williams at Harvard Medical School trained monkeys to play a version of the prisoner's dilemma, a game used to study cooperation. The monkeys sat next to each other and decided whether or not they wanted to cooperate with their companion, by moving a joystick to pick either option. Moving the joystick towards an orange circle meant cooperate, a blue triangle meant "not this time". Neither monkey could see the other's face, or receive any clues about their planned action. If the monkeys cooperated, both received four drops of juice. If one cooperated and the other decided not to, the one who cooperated received one drop, and the other received six drops of juice. If both declined to work together they both received two drops of juice. Once both had made their selections, they could see what the other monkey had chosen and hear the amount of juice their companion was enjoying. © Copyright Reed Business Information Ltd.
Link ID: 20627 - Posted: 02.27.2015
Elizabeth Gibney DeepMind, the Google-owned artificial-intelligence company, has revealed how it created a single computer algorithm that can learn how to play 49 different arcade games, including the 1970s classics Pong and Space Invaders. In more than half of those games, the computer became skilled enough to beat a professional human player. The algorithm — which has generated a buzz since publication of a preliminary version in 2013 (V. Mnih et al. Preprint at http://arxiv.org/abs/1312.5602; 2013) — is the first artificial-intelligence (AI) system that can learn a variety of tasks from scratch given only the same, minimal starting information. “The fact that you have one system that can learn several games, without any tweaking from game to game, is surprising and pretty impressive,” says Nathan Sprague, a machine-learning scientist at James Madison University in Harrisonburg, Virginia. DeepMind, which is based in London, says that the brain-inspired system could also provide insights into human intelligence. “Neuroscientists are studying intelligence and decision-making, and here’s a very clean test bed for those ideas,” says Demis Hassabis, co-founder of DeepMind. He and his colleagues describe the gaming algorithm in a paper published this week (V. Mnih et al. Nature 518, 529–533; 2015. Games are to AI researchers what fruit flies are to biology — a stripped-back system in which to test theories, says Richard Sutton, a computer scientist who studies reinforcement learning at the University of Alberta in Edmonton, Canada. “Understanding the mind is an incredibly difficult problem, but games allow you to break it down into parts that you can study,” he says. But so far, most human-beating computers — such as IBM’s Deep Blue, which beat chess world champion Garry Kasparov in 1997, and the recently unveiled algorithm that plays Texas Hold ’Em poker essentially perfectly (see Nature http://doi.org/2dw; 2015)—excel at only one game. © 2015 Nature Publishing Group
People with attention deficit hyperactivity disorder are about twice as likely to die prematurely as those without the disorder, say researchers. Researchers followed 1.92 million Danes, including 32,000 with ADHD, from birth through to 2013. "In this nationwide prospective cohort study with up to 32-year followup, children, adolescents and adults with ADHD had decreased life expectancy and more than double the risk of death compared with people without ADHD," Soren Dalsgaard, from Aarhus University in Denmark, and his co-authors concluded in Wednesday's online issue of Lancet. Actress Kirstie Alley holds a picture of Raymond Perone while testifying in favour of a bill designed to curb the over-prescribing of psychotropic drugs. Danish researchers studying ADHD say medications can reduce symptoms of inattention and impulsivity. (Phil Coale/Associated Press) "People diagnosed with ADHD in adulthood had a greater risk of death than did those diagnosed in childhood and adolescence. This finding could be caused by persistent ADHD being a more severe form of the disorder." Of the 107 individuals with ADHD who died, information on cause of death was available for 79. Of those, 25 died from natural causes and 54 from unnatural causes, including 42 from accidents. Being diagnosed with ADHD along with oppositional defiant disorder, conduct disorder and substance use disorder also increased the risk of death, the researchers found. Mortality risk was also higher for females than males, which led Dalsgaard to stress the need for early diagnosis, especially in girls and women, and to treat co-existing disorders. Although talk of premature death will worry parents and patients, they can seek solace in knowing the absolute risk of premature death at an individual level is low and can be greatly reduced with treatment, Stephen Faraone, a professor of psychiatry and director of child and adolescent psychiatry research at SUNY Upstate Medical University in New York, said in a journal commentary published with the study. ©2015 CBC/Radio-Canada.
|By Matthew Hutson We like to think of our moral judgments as consistent, but they can be as capricious as moods. Research reveals that such judgments are swayed by incidental emotions and perceptions—for instance, people become more moralistic when they feel dirty or sense contamination, such as in the presence of moldy food. Now a series of studies shows that hippies, the obese and “trailer trash” suffer prejudicial treatment because they tend to elicit disgust. Researchers asked volunteers to read short paragraphs about people committing what many consider to be impure acts, such as watching pornography, swearing or being messy. Some of the paragraphs described the individuals as being a hippie, obese or trailer trash—and the volunteers judged these fictional sinners more harshly, according to the paper in the Journal of Experimental Psychology: General. Questionnaires revealed that feelings of disgust toward these groups were driving the volunteers' assessments. A series of follow-up studies solidified the link, finding that these groups also garnered greater praise for purity-related virtues, such as keeping a neat cubicle. If the transgression in question did not involve purity, such as not tipping a waiter, the difference in judgment disappeared. “The assumption people have is that we draw on values that are universal and important,” says social psychologist E. J. Masicampo of Wake Forest University, who led the study, “but something like mentioning that a person is overweight can really push that judgment around. It's triggering these gut-level emotions.” The researchers also looked for real-world effects. © 2015 Scientific American