Links for Keyword: Attention
Follow us on Facebook and Twitter, or subscribe to our mailing list, to receive news updates. Learn more.
By Daisy Grewal How good are you at multi-tasking? The way you answer that question may tell you more than you think. According to recent research, the better people think they are at multitasking, the worse they actually are at it. And the more that you think you are good at it, the more likely you are to multi-task when driving. Maybe the problem of distracted driving has less to do with the widespread use of smartphones and more to do with our inability to recognize our own limits. A study by David Sanbonmatsu and his colleagues looked at the relationship between people’s beliefs about their own multi-tasking ability and their likelihood of using a cell phone when driving. Importantly, the study also measured people’s actual multi-tasking abilities. The researchers found that people who thought they were good at multi-tasking were actually the worst at it. They were also the most likely to report frequently using their cell phones when driving. This may help explain why warning people about the dangers of cell phone use when driving hasn’t done much to curb the behavior. The study is another reminder that we are surprisingly poor judges of our own abilities. Research has found that people overestimate their own qualities in a number of areas including intelligence, physical health, and popularity. Furthermore, the worse we are at something, the more likely we may be to judge ourselves as competent at it. Psychologists David Dunning and Justin Kruger have studied how incompetence, ironically, is often the result of not being able to accurately judge one’s own incompetence. In one study, they found that people who scored the lowest on tests of grammar and logic were the most likely to overestimate their own abilities. The reverse was also true: the more competent people were most likely to underestimate their abilities. And multi-tasking may be just yet another area where incompetence breeds over-confidence. © 2013 Scientific American
Why do some people feel as though one of their body parts is not truly part of them and go to crazy lengths to get rid of it? Paul D. McGeochanswers: Certain people hold a deep desire to amputate a healthy limb. They are not psychotic, and they fully realize that what they want is abnormal. Nevertheless, they have felt from childhood that the presence of a specific limb, usually a leg, somehow makes their body “overcomplete.” Ultimately, many will achieve their desired amputation through self-inflicted damage or surgery. During the past few years my work with neuroscientists Vilayanur S. Ramachandran of U.C.S.D. and David Brang of Northwestern University, along with research by neuroscientist Peter Brugger of University Hospital Zurich in Switzerland, has transformed our understanding of this condition. Our findings suggest that a dysfunction of specific brain areas on the right side of the brain, which are involved in generating our body image, may explain the desire. Bizarre disorders of body image have long been known to arise after a stroke or other incident inflicts damage to the right side of the brain, particularly in the parietal lobe. The right posterior parietal cortex seems to combine several incoming streams of information—touch, joint position sense, vision and balance—to form a dynamic body image that changes as we interact with the world around us. In brain scans, we have found this exact part of the right parietal lobe to activate abnormally in individuals desiring limb removal. Because the primary sensory areas of the brain still function normally, sufferers are able to see and feel the limb in question. Yet they do not experience it as part of their body because the right posterior parietal lobe fails to adequately represent it. The mismatch between a person's actual physical body and his or her body image seems to cause ongoing arousal in the sympathetic nervous system, which may intensify the desire to remove the limb. Given that sufferers date these feelings to childhood, the right parietal dysfunction most likely is congenital or arises in early development. © 2013 Scientific American
Katherine Harmon Courage An infant's innate sense for numbers predicts how their mathematical aptitude will develop years later, a team of US researchers has found. Babies can spot if a set of objects increases or decreases in number — for instance, if the number of dots on a screen grows, even when dot size, colour and arrangement also change. But until recently, researchers could generally only determine the number sense of groups of babies, thus ruling out the ability to correlate this with later mathematics skills in individuals. In 2010, Elizabeth Brannon, a neuroscientist at Duke University in Durham, North Carolina, and her colleagues demonstrated that they could test and track infants' number sense over time1. To do this, six-month-old babies are presented with two screens. One shows a constant number of dots, such as eight, changing in appearance, and the other also shows changing dots but presents different numbers of them — eight sometimes and 16 other times, for instance. An infant who has a good primitive number sense will spend more time gazing at the screen that presents the changing number of dots. In the latest work, which is published in this week's Proceedings of the National Academy of Sciences2, Brannon's team took a group of 48 children who had been tested at six months of age and retested them three years later, using the same dot test but also other standard maths tests for preschoolers — including some that assessed the ability to count, to tell which of two numbers is larger and to do basic calculations. © 2013 Nature Publishing Group
Related chapters from BP7e: Chapter 7: Life-Span Development of the Brain and Behavior; Chapter 18: Attention and Higher Cognition
Related chapters from MM:Chapter 13: Memory, Learning, and Development; Chapter 14: Attention and Consciousness
Link ID: 18822 - Posted: 10.22.2013
by Helen Thomson ONE moment you are alive. The next you are dead. A few hours later and you are alive again. Pharmacologists have discovered a mechanism that triggers Cotard's syndrome – the mysterious condition that leaves people feeling like they, or parts of their body, no longer exist. With the ability to switch the so-called walking corpse syndrome on and off comes the prospect of new insights into how conscious experiences are constructed. Acyclovir – also known by the brand name Zovirax – is a common drug used to treat cold sores and other herpes infections. It usually has no harmful side effects. However, about 1 per cent of people who take the drug orally or intravenously experience some psychiatric side effects, including Cotard's. These occur mainly in people who have renal failure. To investigate the potential link between acyclovir and Cotard's, Anders Helldén at Karolinska University Hospital in Stockholm and Thomas Lindén at the Sahlgrenska Academy in Gothenburg pooled data from Swedish drug databases along with hospital admissions. They identified eight people with acyclovir-induced Cotard's. One woman with renal failure began using acyclovir to treat shingles. She ran into a hospital screaming, says Helldén. After an hour of dialysis, she started to talk: she said the reason she was so anxious was that she had a strong feeling she was dead. After a few more hours of dialysis she said, "I'm not quite sure whether I'm dead any more but I'm still feeling very strange." Four hours later: "I'm pretty sure I'm not dead any more but my left arm is definitely not mine." Within 24 hours, the symptoms had disappeared. © Copyright Reed Business Information Ltd.
by Colin Barras A part of all of us loves sums. Eavesdropping on the brain while people go about their daily activity has revealed the first brain cells specialised for numbers. Josef Parvizi and his colleagues at Stanford University in California enlisted the help of three people with epilepsy whose therapy involved placing a grid of electrodes on the surface of their brain that record activity. Neurons fired in a region called the intraparietal sulcus when the three volunteers performed arithmetic tests, suggesting they dealt with numbers. The team continued to monitor brain activity while the volunteers went about their normal activity in hospital. Comparing video footage of their stay with their brain activity (see video, above) revealed that the neurons remained virtually silent for most of the time, bursting into life only when the volunteers talked about numbers or numerical concepts such as "more than" or "less than". There is debate over whether some neural populations perform many functions or are involved in very precise tasks. "We show here that there is specialisation for numeracy," says Parvizi. Journal reference: Nature Communications, DOI: 10.1038/ncomms3528 © Copyright Reed Business Information Ltd.
Many people, I've heard talk, wonder what's going on inside Republican speaker John Boehner's brain. For cognitive neuroscientists, Boehner's brain is a case study. At the same time, others are frustrated with Democrat Harry Reid. The Senate Majority leader needs to take a tip from our founding fathers. Many of the intellectual giants who founded our democracy were both statesmen and scientists, and they applied the latest in scientific knowledge of their day to advantage in governing. The acoustics of the House of Representatives, now Statuary Hall, allowed John Quincy Adams and his comrades to eavesdrop on other members of congress conversing in whispers on the opposite side of the parabolic-shaped room. Senator Reid, in stark contrast, is still applying ancient techniques used when senators wore togas -- reason and argument -- and we all know how badly that turned out. The search for a path to compromise can be found in the latest research on the neurobiological basis of social behavior. Consider this new finding just published in the journal Brain Research. Oxytocin, a peptide produced in the hypothalamus of the brain and known to cement the strong bond between mother and child at birth, has been found to promote compromise in rivaling groups! This new research suggests that Congresswoman Nancy Pelosi could single-handedly end the Washington deadlock by spritzing a bit of oxytocin in her perfume and wafting it throughout the halls of congress. One can only imagine the loving effect this hormone would have on Senate Republican Ted Cruz, suddenly overwhelmed with an irresistible urge to bond with his colleagues, fawning for a cozy embrace like a babe cuddling in its mother's arms. And it is so simple! No stealthy spiking the opponent's coffee (or third martini at lunch) would be required, oxytocin works when it is inhaled through the nasal passages as an odorless vapor. © 2013 TheHuffingtonPost.com, Inc.
Related chapters from BP7e: Chapter 18: Attention and Higher Cognition; Chapter 1: Biological Psychology: Scope and Outlook
Related chapters from MM:Chapter 14: Attention and Consciousness; Chapter 1: An Introduction to Brain and Behavior
Link ID: 18761 - Posted: 10.08.2013
By Bruce Bower Cartoon ghosts have scared up evidence that the ability to visualize objects in one’s mind materializes between ages 3 and 5. When asked to pick which of two mirror-image ghost cutouts or drawings fit in a ghost-shaped hole, few 3-year-olds, a substantial minority of 4-year-olds and most 5-year-olds regularly succeeded, say psychologist Andrea Frick of the University of Bern, Switzerland, and her colleagues. Girls performed as well as boys on the task, suggesting that men’s much-studied advantage over women in mental rotation doesn’t emerge until after age 5, the researchers report Sept. 17 in Cognitive Development. Mental rotation is a spatial skill regarded as essential for science and math achievement. Most tasks that researchers use to assess mental rotation skills involve pressing keys to indicate whether block patterns oriented at different angles are the same or different. That challenge overwhelms most preschoolers. Babies apparently distinguish block patterns from mirror images of those patterns (SN: 12/20/08, p. 8), but it’s unclear whether that ability enables mental rotation later in life. Frick’s team studied 20 children at each of three ages, with equal numbers of girls and boys. Youngsters saw two ghosts cut out of foam, each a mirror image of the other. Kids were asked to turn the ghosts in their heads and choose the one that would fit like a puzzle piece into a ghost’s outline on a board. Over seven trials, the ghosts were tilted at angles varying from the position of the outline. The researchers used three pairs of ghost cutouts, for a total of 21 trials. © Society for Science & the Public 2000 - 2013
Related chapters from BP7e: Chapter 18: Attention and Higher Cognition; Chapter 12: Sex: Evolutionary, Hormonal, and Neural Bases
Related chapters from MM:Chapter 14: Attention and Consciousness; Chapter 8: Hormones and Sex
Link ID: 18706 - Posted: 09.26.2013
By Neuroskeptic Neuroscientists are interested in how brains interact socially. One of the main topics of study is ‘mentalizing’ aka ‘theory of mind’, the ability to accurately attribute mental states – such as beliefs and emotions – to other people. It is widely believed that the brain has specific areas for this – i.e. social “modules” (although today most neuroscientists are shy about using that word, it’s basically what’s at issue.) But two new papers out this week suggest that people can still mentalize successfully after damage to “key parts of the theory of mind network”. Herbet et al, writing in Cortex, showed few effects of surgical removal of the right frontal lobe in 10 brain tumour patients. On two different mentalizing tasks, they showed that removal caused either no decline in performance, or only a transient one. Meanwhile Michel et al report that the left temporal pole is dispensable for mentalizing as well, in a single case report in the Journal of Cognitive Neuroscience. They describe a patient suffering from frontotemporal dementia (FTD), whose left temporal lobe was severely atrophied. He’d lost the use of language, but he did quite normally on theory of mind tests adapted to be non-linguistic. In both papers, these patients don’t have those parts of the brain that are most activated in fMRI studies of mentalizing. Where the blobs on the brain normally go, they have no brain.
By Melissa Hogenboom Science reporter, BBC News Smaller animals tend to perceive time in slow-motion, a new study has shown. This means that they can observe movement on a finer timescale than bigger creatures, allowing them to escape from larger predators. Insects and small birds, for example, can see more information in one second than a larger animal such as an elephant. The work is published in the journal Animal Behaviour. "The ability to perceive time on very small scales may be the difference between life and death for fast-moving organisms such as predators and their prey," said lead author Kevin Healy, at Trinity College Dublin (TCD), Ireland. The reverse was found in bigger animals which may miss things that smaller creatures can rapidly spot. In humans, too, there is variation among individuals. Athletes, for example, can often process visual information more quickly. An experienced goalkeeper would therefore be quicker than others in observing where a ball comes from. The speed at which humans absorb visual information is also age-related, said Andrew Jackson, a co-author of the work at TCD. "Younger people can react more quickly than older people, and this ability falls off further with increasing age." The team looked at the variation of time perception across a variety of animals. They gathered datasets from other teams who had used a technique called critical flicker fusion frequency, which measures the speed at which the eye can process light. BBC © 2013
Related chapters from BP7e: Chapter 18: Attention and Higher Cognition; Chapter 10: Vision: From Eye to Brain
Related chapters from MM:Chapter 14: Attention and Consciousness; Chapter 7: Vision: From Eye to Brain
Link ID: 18651 - Posted: 09.16.2013
By Josh Shaffer DURHAM It’s not often that the high-minded world of neuroscience collides with the corny, old-fashioned art of ventriloquism. One depends on dummies; the other excludes them. But a Duke University study uses puppet-based comedy to demonstrate the complicated inner workings of the brain and shows what every ventriloquist knows: The eye is more convincing than the ear. The study, which appears in the journal PLOS ONE, seeks to explain how the brain combines information coming from two different senses. How, asks Duke psychology and neuroscience professor Jennifer Groh, does the brain determine where a sound is coming from? In your eyes, the retina takes a snapshot, she said. It makes a topographic image of what’s in front of you. But the ears have nothing concrete to go on. They have to rely on how loud the sound is, how far away and from what direction. That’s where a ventriloquist comes in, providing a model for this problem. With a puppet, the noise and the movement are coming from different places. So how does the brain fix this and choose where to look? Duke researchers tested their hypotheses on 11 people and two monkeys, placing them in a soundproof booth.
Related chapters from BP7e: Chapter 18: Attention and Higher Cognition; Chapter 9: Hearing, Vestibular Perception, Taste, and Smell
Related chapters from MM:Chapter 14: Attention and Consciousness; Chapter 6: Hearing, Balance, Taste, and Smell
Link ID: 18619 - Posted: 09.09.2013
by Colin Barras Familiarity may breed contempt, and it also makes it easier to ignore our nearest and dearest. The human brain has an uncanny ability to focus on one voice in a sea of chatterSpeaker, for example, at a party, but exactly how it does so is still up for debate. "In the past, people have looked at the acoustic characteristics that enable the brain to do this," says Ingrid Johnsrude at Queen's University in Kingston, Ontario, Canada. "Things like differences in voice pitch or its timbre." Johnsrude and her colleagues wondered if the familiarity of the voice also plays a role. Can people focus on one voice in a crowd more effectively if it belongs to a close relation? And is a familiar voice more easily ignored if we want to listen to someone else? To find out, the team recruited 23 married couples. Each had been married and living together for at least 18 years. Individuals were played two sentences simultaneously and asked to report back details about one of them, such as the colour and number mentioned. They did this correctly 80 per cent of the time when their spouse spoke the target sentence and a stranger spoke the decoy sentence. If strangers spoke both, the success rate dropped to 65 per cent. © Copyright Reed Business Information Ltd
Related chapters from BP7e: Chapter 18: Attention and Higher Cognition; Chapter 9: Hearing, Vestibular Perception, Taste, and Smell
Related chapters from MM:Chapter 14: Attention and Consciousness; Chapter 6: Hearing, Balance, Taste, and Smell
Link ID: 18584 - Posted: 08.31.2013
Elizabeth Norton Brain cells, like Henry Higgins in My Fair Lady, grow accustomed to a familiar face—so much so that repeatedly viewing a distorted face will make the normal face look odd. This process, known as visual adaptation, is enhanced by sleep and may be an essential component of memory, a new study finds. After multiple exposures to a striking visual pattern, neurons in the retina and visual cortex of the brain fire less frequently the next time you see the pattern. By devoting less energy to familiar sights, the brain is free to concentrate on the next new thing that comes along; the original image becomes a routine perception. Scientists think that this allocation of mental resources is crucial to our ability to perceive and interpret our surroundings. Whether visual adaptation is a prelude to memory formation is another question, one that intrigued cognitive neuroscientist Thomas Ditye of University College London. Because sleep strengthens memory, Ditye and colleagues decided to test whether visual adaptation also improves after some shuteye. The researchers asked a group to view a computer screen on which distorted images of the faces of actors George Clooney and Angelina Jolie flashed for periods of 0.5 to 6 seconds. The images were “extended”—stretched until they achieved the blown-up look of a fun house mirror. The object of the test was to determine whether the brain would adapt to images and begin seeing the distorted faces as normal. The volunteers, however, believing their reaction time was being tested, merely pressed a button whenever they saw the image. © 2012 American Association for the Advancement of Science
By Susan Gaidos If you’re someone who enjoys being recognized, Julian Lim is your kind of waiter. Lim, who’s working his way through college waiting tables, remembers the face of everyone that walks through the door of the South Bend, Ind., restaurant where he works. His abilities go beyond making his customers feel special. This spring, when he cut his hand on broken glass, he pegged the emergency room nurse as a fellow student from his grade school days. Though they’d never spoken, and the girl had since undergone changes in appearance, Lim recognized her instantly. Carrie Shanafelt is good with faces, too. A professor of literature at Grinnell College in Iowa, Shanafelt can spot her students outside the classroom, whether it’s the first week of class or years later. And Ajay Jansari, an information technology specialist in London, often has to see a face only once to remember it, even those he meets thousands of miles from home. While some people say they never forget a face, these folks have scientific studies to back their claims. Called “super recognizers,” they’re among a small group of individuals being studied by scientists at Dartmouth College and in England to better understand how some people can recognize almost every face they have ever seen. Scientists are now putting super recognizers’ skills to the test to get a handle on how face-processing areas of the brain work to make a few people so adept at recalling faces. Findings from the studies may advance understanding of how most people categorize faces — a subject that is still poorly understood. © Society for Science & the Public 2000 - 2013
By Jessica Shugart People who need sugary snacks to stay sharp throughout the day could be prisoners of their own beliefs. The brain works just fine without regular shots of sugar in people who believe their willpower is unlimited, a new study shows. “There's a dominant theory in psychology that willpower is limited, and whenever you exert yourself to do a hard task or to resist a temptation, you deplete this limited resource,” says psychologist Carol Dweck from Stanford University. Previous studies have shown that mental exertion diminishes blood glucose levels and that a person’s willpower can be rejuvenated by ingesting a sugary drink. But Dweck’s earlier work led her to suspect that people’s attitudes about willpower may be responsible for that effect. In the new study, published online August 19 in the Proceedings of the National Academy of Sciences, Dweck, along with colleagues at the University of Zurich in Switzerland, focused on how attitudes about willpower may shape a person’s sugar dependence in the face of a challenge. The scientists also tested whether altering these beliefs might liberate a person from such a calorie-rich requirement. In the first of three experiments, the researchers asked students about their attitudes on willpower, then gave them lemonade sweetened with either sugar or a sugar substitute. Ten minutes after downing the sweet beverage, the students took tests of self-control and mental acuity. The students who subscribed to a self-generating belief about unlimited willpower scored equally well whether their drinks contained sugar or not. But the students who felt willpower was limited needed sugar to perform as well as the other group did. © Society for Science & the Public 2000 - 2013
Related chapters from BP7e: Chapter 13: Homeostasis: Active Regulation of the Internal Environment; Chapter 18: Attention and Higher Cognition
Related chapters from MM:Chapter 9: Homeostasis: Active Regulation of the Internal Environment; Chapter 14: Attention and Consciousness
Link ID: 18537 - Posted: 08.20.2013
What people experience as death creeps in—after the heart stops and the brain becomes starved of oxygen—seems to lie beyond the reach of science. But the authors of a new study on dying rats make a bold claim: After cardiac arrest, the rodents’ brains enter a state similar to heightened consciousness in humans. The researchers suggest that if the same is true for people, such brain activity could be the source of the visions and other sensations that make up so-called near-death experiences. Estimated to occur in about 20% of patients who survive cardiac arrest, near-death experiences are frequently described as hypervivid or “realer-than-real,” and often include leaving the body and observing oneself from outside, or seeing a bright light. The similarities between these reports are hard to ignore, but the conversation about near-death experiences often bleeds into metaphysics: Are these visions produced solely by the brain, or are they a glimpse at an afterlife outside the body? Neurologist Jimo Borjigin of the University of Michigan, Ann Arbor, got interested in near-death experiences during a different project—measuring the hormone levels in the brains of rodents after a stroke. Some of the animals in her lab died unexpectedly, and her measurements captured a surge in neurochemicals at the moment of their death. Previous research in rodents and humans has shown that electrical activity surges in the brain right after the heart stops, then goes flat after a few seconds. Without any evidence that this final blip contains meaningful brain activity, Borjigin says “it’s perhaps natural for people to assume that [near-death] experiences came from elsewhere, from more supernatural sources.” But after seeing those neurochemical surges in her animals, she wondered about those last few seconds, hypothesizing that even experiences seeming to stretch for days in a person’s memory could originate from a brief “knee-jerk reaction” of the dying brain. © 2012 American Association for the Advancement of Science.
By Nathan Seppa The late rock and roll singer Jim Morrison was not a poster boy for public safety — and was no authority on safe driving. After all, later in “Roadhouse Blues,” he has beer for breakfast. But the opening line of that Doors’ song still resonates as sound guidance. If only such good advice could stand the test of time. “Roadhouse Blues” hit the airwaves in 1970, long before the unlikely marriage of driving and talking on a cell phone. Millions of people now routinely conduct remote conversations while driving, despite research showing that it’s dangerous — even with two eyes on the road and both hands upon the wheel. It turns out that hands don’t matter. It’s the conversation that can be lethal. Cell phone conversations impede what a driver sees and processes, a number of studies have shown. That, in turn, slows reactions and other faculties. This distracted state should be familiar to everyone. “That’s why you can drive home and not remember having driven home,” says Daniel Simons, a psychologist at the University of Illinois at Urbana-Champaign. “Just because you look at something doesn’t mean you see it.” Simons has shown that people assigned to observe certain activities in a lab setting can totally miss other events occurring in the very same space. The on-road versions of such blind spots show up when drivers engaged in a cell phone conversation fail to look at side streets or watch for pedestrians. This distraction may seem subtle and even fleeting, but it takes a toll: The risk of an accident quadruples when the driver is on the phone, studies have suggested. © Society for Science & the Public 2000 - 2013
Keith Barry, skilled magician that he is, doesn't give away his tricks but he does give the audience a clue when he says "magic is all about directing attention." Neuroscientists have long known that attention plays a key role in perception, and yet, we still don't fully understand the details of how attention works and what neural mechanisms are involved. Only a small fraction of the information that comes in through our eyes is actually perceived by our conscious brains. Attention is the filter that directs what is most salient in our environment to our conscious awareness. Almost all magic tricks somehow take advantage of loopholes in attention. For instance, a key strategy for magicians taps into something which cognitive neuroscientists call "inattentional or perceptual blindness", our inability to notice an object or feature in a visual scene because attention is directed elsewhere. You have experienced this phenomenon yourself. Your brain is constantly bombarded with stimuli, and it is impossible to pay attention to them all. While your attention is focused on one thing -- neuroscientists call this the "attentional spotlight"-- your ability to perceive objects outside this focus area is compromised. Indeed, if we could record the activity of the neurons in your brain that track to the visual scene, the neural responses for those areas outside the attentional spotlight would be dampened. The magician takes advantage of this phenomenon. By distracting your attention with sly hand movements, lively banter, humor, or skillful shifts of gaze, he can move your "attentional spotlight," while manipulating the action elsewhere, all without your knowing it and indeed while you think you are paying close attention! So, there you have it -- the neuroscientific answer to "how did he do that?" But, the fact that there is a logical, brain-based explanation behind the magician's tricks is not so surprising. The more interesting question is, if neuroscience can explain magic, can magic teach us anything about neuroscience? © 2013 TheHuffingtonPost.com, Inc.
By Lucas Laursen My cousin Guillermo Cassinello Toscano was on the train that derailed in Santiago de Compostela, Spain, last week when it went around a bend at twice the speed limit. Cassinello heard a loud vibration and then a powerful bump and then found himself surrounded by bloody bodies in wagon number nine. Shaking, he escaped the wreckage through either a door or a hole in the train—he cannot recall—then sat amid the smoke and debris next to the track and began to cry. Seventy-nine passengers died. Cassinello doesn’t remember everything that happened to him. The same mechanisms that kept his brain sharp enough to escape immediate danger may also make it harder for him both to recall the accident, and to put the trauma behind him. "The normal thing is that the person doesn't remember the moment of the accident or right after," says clinical psychologist Javier Rodriguez Escobar of trauma therapy team Grupo Isis in Seville, who helped treat and study victims of the 2004 Madrid train bombings. That's because the mind and the body enter a more alert but also more stressed state, with trade-offs that can save your life, but harm your mind’s memory-making abilities. As the train fell over, several changes would have swept through Cassinello’s body. His adrenal glands, near his kidneys, would have released adrenaline (also known as epinephrine) into his bloodstream. The adrenaline would have directed blood to the powerful muscles of his arms and legs, where it would help him escape the wreckage faster. The hormone would have raised his heart and breathing rates. It also would have stimulated his vagus nerve, which runs from his spine to his brain. Although adrenaline cannot cross the blood–brain barrier, the vagus can promote noradrenaline production in the brain. That hormone activates the amygdala, which helps form memories. © 2013 Scientific American
By Jessica Shugart Lunch at a restaurant with a friend could lessen the brain’s aptitude for detailed tasks back at work, a new study suggests. If an error-free afternoon is the goal, perhaps workers should consider hastily consuming calories alone at their desks. But bosses shouldn’t rush to glue workers to their chairs just yet. The research is only a first stab at teasing out how a sociable lunch affects work performance, says study leader Werner Sommer of Humboldt University in Berlin. Researchers have long thought that dining with others fosters mental well-being, cooperation and creativity. To test the effects of a midday social hour on the brain’s capacity to get through the workday, Sommer and his colleagues gave 32 university students lunch in one of two settings and then tested their mental focus. Half of the students enjoyed meals over a leisurely hour with a friend at a casual Italian restaurant. The other group picked up their meals from the same restaurant, but had only 20 minutes to eat alone in a drab office. People who went out to lunch got to choose from a limited vegetarian menu; participants in the office group had meals that matched the choice of a member of the other group. After lunch, the group that dined in bland solitude performed better on a task that assesses rapid decision making and focus, the researchers report July 31 in PLOS ONE. Measurements of brain activity also suggested that the brain’s error-monitoring system could be running at sub-par levels in those who ate out. © Society for Science & the Public 2000 - 2013
Avoiding temptation works better than relying on willpower alone, a study of brain activity finds. "Struggles with self-control pervade daily life and characterize an array of dysfunctional behaviours, including addiction, overeating, overspending and procrastination," Molly Crockett, a postdoctoral fellow at University College London, and her co-authors said in today's issue of the journal Neuron. "Our research suggests that the most effective way to beat temptations is to avoid facing them in the first place," she said in a release. In the experiment, researchers studied 58 healthy heterosexual males in Cambridge and 20 in Amsterdam. Investigators used functional MRI as part of the study of self-control to explore the neural mechanisms involved. At the beginning of the trial, participants were shown a series of 400 images of women in lingerie or swimwear and were asked to rank them on a scale of zero to 10 on how enjoyable they were. Each man's preferences were then used to present small, short-term rewards or a large reward after a delay. Small rewards were mildly enjoyable erotic pictures and large rewards were extremely enjoyable ones. (The scientists said they could not use money, for example, since subjects could only reap the rewards of money once they left the lab. Food rewards like juice could interfere with the MRI readings.) © CBC 2013
Related chapters from BP7e: Chapter 18: Attention and Higher Cognition; Chapter 13: Homeostasis: Active Regulation of the Internal Environment
Related chapters from MM:Chapter 14: Attention and Consciousness; Chapter 9: Homeostasis: Active Regulation of the Internal Environment
Link ID: 18405 - Posted: 07.25.2013