Chapter 17. Learning and Memory
Follow us on Facebook and Twitter, or subscribe to our mailing list, to receive news updates. Learn more.
A brain-training video game that improved the vision of college baseball players by as much as two lines on an eye chart has been developed by U.S. researchers. "This is something which I think could help almost anybody," said Aaron Seitz, a neuroscientist at the University of California, Riverside, who the led the research. Players on the university's baseball team improved their visual acuity by 31 per cent after training with the app. And that translated into better performance on the baseball field, where better vision improves the odds of hitting a ball travelling well over 100 km/h. "What we found is they had fewer strikeouts, they were able to create more runs," Seitz told CBC's Quirks & Quarks in an interview that airs Saturday. The players had more runs than predicted even after taking into account the natural improvement that would be expected over the course of the season. Further calculations suggest the improved performance helped the team to win four or five additional games. Following 30 sessions of training with the app, players had better vision, fewer strikeouts, more runs and more wins. But Seitz thinks the app has even more potential to help people with eye conditions such as lazy eye, glaucoma, or age-related macular degeneration. There are 100 million people around the world who have such low vision that glasses don't help, he added. "All that they have to gain is the brain training element.… For these people, there's just really big real-world benefits that could be achieved if we're able to improve their vision."
|By Beth Skwarecki Prions, the protein family notorious for causing "mad cow" and neurodegenerative diseases like Parkinson's, can play an important role in healthy cells. "Do you think God created prions just to kill?" mused Nobel laureate Eric Kandel. "These things must have evolved initially to have a physiological function." His work on memory helped reveal that animals make and use prions in their nervous systems as part of an essential function: stabilizing the synapses that constitute long-term memories. These natural prions aren't infectious but on a molecular level they chain up exactly the same way as their disease-causing brethren. (Some researchers call them "prionlike" to avoid confusion.) This week, work from neuroscientist Kausik Si of the Stowers Institute for Medical Research, one of Kandel's former students, shows that the prion's action is tightly controlled by the cell, and can be turned on when a new long-term memory needs to be formed. Prions are proteins with two unusual properties: First, they can switch between two possible shapes, one that is stable on its own and an alternate conformation that can form chains. Second, the chain-forming version has to be able to trigger others to change shape and join the chain. Say that in the normal version the protein is folded so that one portion of the protein structure—call it "tab A"—fits into its own "slot B." In the alternate form, though, tab A is available to fit into its neighbor's slot B. That means the neighbor can do the same thing to the next protein to come along, forming a chain or clump that can grow indefinitely. © 2014 Scientific American,
James Hamblin Brain training is becoming big business. Everywhere you look, someone is talking about neuroplasticity and trying to train your brain. Soon there will be no wild brains left. At the same time, everyone who spends more than two continuous hours using a computer is, according to the American Optometric Association, ruining their eyes with Computer Vision Syndrome. So, Dr. Aaron Seitz might be onto something with his new brain-training program that promises better vision. UltimEyes is a game-based app that's sold as "fun and rewarding" as it improves your vision and "reverse[s] the effects of aging eyes." It doesn't claim to work on the eyes themselves, but on the brain cortex that processes vision—the part that takes blurry puzzle pieces from the eyes and arranges them into a sweet puzzle. (Brain training for memory, the kind we hear about the most on TV, would be the part that lacquers the finished puzzle, frames it, and hangs it on the wall.) A standard 25-minute session using UltimEyes forces your eyes to work in ways they probably don't in everyday life, and its website warns that after the first use, "just like the first time that you go to the gym, your eyes may feel a bit tired. This experience typically goes away by your third session as your visual system adjusts to its new work-out routine." Seitz is a neuroscientist at the University of California, Riverside. To test out his vision-training game, he had players on the university's baseball team use the app. Half the team trained for 30 sessions. For comparison, the other half did no training. © 2014 by The Atlantic Monthly Group
By GREGORY COWLES David Stuart MacLean’s first book, “The Answer to the Riddle Is Me,” opens with a scene out of Robert Ludlum: The protagonist wakes from a blackout to find himself on a crowded train platform in India, with no idea who he is or what he’s doing in a foreign country. The catch is that the protagonist is Mr. MacLean himself, and his book isn’t an international thriller but a “memoir of amnesia,” as his agreeably paradoxical subtitle puts it — the true story of how his memory was wiped clean and how that condition has subsequently affected his life. It is all the more thrilling for that. In 2002, Mr. MacLean was a 28-year-old Fulbright scholar visiting India to research a novel. It wasn’t his first trip; he had gone a few years earlier and stayed for months. But this time around, his anti-malaria medication touched off a break with reality as sudden as it was severe. He hallucinated angels and demons, and felt his thoughts “puddling in the carpet near the doorway and sloshing down the hall.” Delirious, he agreed with the police officer who surmised he must be a drug addict, and apologized profusely for misdeeds he had never committed. At the hospital, a nurse called him “the most entertaining psychotic that they’d ever had.” As harrowing as this territory is, Mr. MacLean makes an affable, sure-footed guide. In his descriptions, you can recognize the good fiction writer he must have been even before amnesia forced him to view the world anew; if the writer’s task is to “make it new,” then losing your memory turns out to be an unexpected boon. An avid drinker before his breakdown, he recoils the first time he tries Scotch again, thinking it smells “like Band-Aids.” He can’t remember his girlfriend of a year, but her voice is “faintly familiar, like the smell of the car heater the first time you turn it on in the fall.” He grasps at hope when his parents arrive to take him home: “I still didn’t have my memory, but I now had an outline of myself, like a tin form waiting for batter.” © 2014 The New York Times Company
Keyword: Learning & Memory
Link ID: 19262 - Posted: 02.18.2014
Katherine Sharpe Ben Harkless could not sit still. At home, the athletic ten-year-old preferred doing three activities at once: playing with his iPad, say, while watching television and rolling on an exercise ball. Sometimes he kicked the walls; other times, he literally bounced off them. School was another story, however. Ben sat in class most days with his head down on his desk, “a defeated heap”, remembers his mother, Suzanne Harkless, a social worker in Berkeley, California. His grades were poor, and his teacher was at a loss for what to do. Harkless took Ben to a therapist who diagnosed him with attention deficit hyperactivity disorder (ADHD). He was prescribed methylphenidate, a stimulant used to improve focus in people with the condition. Harkless was reluctant to medicate her child, so she gave him a dose on a morning when she could visit the school to observe. “He didn't whip through his work, but he finished his work,” she says. “And then he went on and helped his classmate next to him. My jaw dropped.” ADHD diagnoses are rising rapidly around the world and especially in the United States, where 11% of children aged between 4 and 17 years old have been diagnosed with the disorder. Between half and two-thirds of those are put on medication, a decision often influenced by a child's difficulties at school. And there are numerous reports of adolescents and young adults without ADHD using the drugs as study aids. © 2014 Nature Publishing Group,
By Roy H. Hamilton and Jihad Zreik It's hard to imagine anyone, no matter how brilliant, who doesn't yearn to be even smarter. Thanks to recent advances in neural science, that wish may come true. Researchers are finding ways to rev up the human brain like never before. There would be just one question: Do we really want to inhabit that world? It may be too late to ask. Modern society has already embraced the basic idea of fine-tuning our intellects via artificial procedures—what might be termed “cosmetic” neurology. Schoolchildren take Adderall, Concerta and other attention-focusing medications. Parents and teachers rely on antidepressants and antianxiety drugs. And self-help books offer the latest advances in neuroscience to help ordinary people think faster and sharper. Add to those advances another cognitive-enhancement method: transcranial direct-current stimulation (tDCS). With this technique, electrodes applied to the scalp deliver minuscule amperages of current to the brain. This trickle of electricity seems to cause incremental adjustments in the electrical potentials of membranes in the neurons closest to the electrodes, increasing or decreasing their likelihood of firing. And that, in turn, induces measurable changes in memory, language, mood, motor function, attention and other cognitive domains. Investigators still aren't sure whether tDCS can cause long-term neural changes. Although most tests show only transient effects, there is limited evidence that repeated applications might have more persistent results. The procedure is not approved by the U.S. Food and Drug Administration, and the consensus among experts is that it should be performed only under qualified supervision. Nevertheless, if used properly, it is safe, portable, easy to implement and inexpensive. © 2014 Scientific American
By Daniel Engber Drop a mouse in some water and white paint, and it will know just what to do. Mice can swim, by whipping their tails like a flagellum, but they don't like doing it; a mouse in a tub tries to find a way out. There's no need for training, or food pellets, or annoying electric shocks: To put a mouse through a water maze, you need only to build a little platform for it, hidden somewhere just beneath the surface. The mouse will try to find that platform without any encouragement. It's a setup that's so simple—and so useful in measuring an animal's capacity for learning and memory—it hardly seems like it would need inventing. But it took a cognitive neuroscientist at the University of St. Andrews in Scotland to come up with the tub-and-platform method. In 1979, Richard Morris built a heated pool about 4 feet and 3 inches in diameter, filled it with water and fresh milk, and then added a platform made of stones and drain piping. Within a few years, his method (designed for rats) had been adapted for smaller lab mice, and had made its way into rodent labs around the world. Now it's among the most widespread animal-testing protocols in all of biomedicine. Scientists plunge mice in murky water to test the effects of brain damage, or the functions of particular genes on learning, or the efficacy of new drugs for treating Alzheimer's. You can even buy a standard-issue "Morris Water Maze" direct from a lab-supply shop, along with specialized software for recording its results. That fact that so few of us would call a tub full of milk a “maze” only goes to show that rodent mazes aren't what they used to be. Early psychologists tempted rats with tricky blind alleys and wrong turns using contraptions built by hand, of wood and wire. © 2014 The Slate Group LLC.
Keyword: Learning & Memory
Link ID: 19233 - Posted: 02.11.2014
Memory can be altered by new experience, and isn't nearly as accurate as courtroom testimony might have us believe, a new study suggests. The results suggest a cheeky answer to the question posed by comedian Richard Pryor: "Who you gonna believe: me, or your lyin' eyes?" Turns out, Pryor was onto something. The brain behind our eyes can distort reality or verify it, based on subsequent experience. And somewhat paradoxically, the same area of the brain appears to be strongly involved in both activities, according to a study published online Tuesday in the Journal of Neuroscience. Northwestern University cognitive neuroscientist Donna Bridge was testing how memory is either consolidated or altered, by giving 17 subjects a deceptively simple task. They studied the location of dozens of objects briefly flashed at varied locations on a standard computer screen, then were asked to recall the object's original location on a new screen with a different background. When subjects were told to use a mouse to drag the re-presented object from the center of the new screen to the place where they recalled it had been located, 16 of 17 got it wrong, by an average of about 3 inches. When the same subjects then were given three choices - the original location, the wrong guess and a neutral spot between them - they almost unfailingly dragged the object to the incorrectly recalled location, regardless of the background screen. Their new memory was false. © 2014 Hearst Communications, Inc.
Keyword: Learning & Memory
Link ID: 19224 - Posted: 02.08.2014
| by Isaac Saul Multi-step puzzles can be difficult for humans, but what if I told you there was a bird that could solve them on its own? In this BBC special, Dr. Alex Taylor has set up an eight-step puzzle to try and stump one of the smartest crows he's seen in captivity. They describe the puzzle as "one of the most complex tests of the animal mind ever." This isn't the first time crows' intelligence has been tested, either. Along with being problem solvers, these animals have an eerie tendency towards complex human-like memory skills. Through several different studies, we've learned that crows can recognize faces, communicate details of an event to each other and even avoid places they recognize as dangerous. This bird, dubbed "007" for its crafty mind, flies into the caged puzzle and spends only seconds analyzing the puzzle before getting down to business. Despite the puzzle's difficulty, the bird only seems to be stumped momentarily. At the end of the puzzle is a food reward, but how he gets there is what will really blow your mind. © 2014 TheHuffingtonPost.com, Inc
Karen Weintraub, Every time you pull up a memory – say of your first kiss – your mind reinterprets it for the present day, new research suggests. If you're in the middle of an ugly divorce, for example, you might recall it differently than if you're happily married and life is going well. This makes your memory quite unlike the video camera you may imagine it to be. But new research in the Journal of Neuroscience suggests it's very effective for helping us adapt to our environments, said co-author Joel Voss, a researcher at Northwestern University's Feinberg School of Medicine. Voss' findings build on others and may also explain why we can be thoroughly convinced that something happened when it didn't, and why eyewitness testimony is notoriously unreliable. The new research also suggests that memory problems like those seen in Alzheimer's could involve a "freezing" of these memories — an inability to adapt the memory to the present, Voss said. Our memories are thus less a snapshot of the past, than "a record of our current view on the past," said Donna Rose Addis, a researcher and associate professor at the University of Auckland in New Zealand, who was not involved in the research. Using brain scans of 17 healthy volunteers as they were taught new data and recalled previously learned information, Voss and his colleagues were able to show for the first time precisely when and where new information gets implanted into existing memories.
Keyword: Learning & Memory
Link ID: 19205 - Posted: 02.05.2014
by Susan Milius Male bee flies fooled into trying to copulate with a daisy may learn from the awkward incident. Certain orchids and several forms of South Africa’s Gorteria diffusa daisy lure pollinators by mimicking female insects. The most effective daisy seducers row a dark, somewhat fly-shaped bump on one of their otherwise yellow-to-orange petals. Males of small, dark Megapalpus capensis bee flies go wild. But tests show the daisy’s victims waste less time trying to mate with a second deceptive daisy than with the first. “Far from being slow and stupid, these males are actually quite keen observers and fairly perceptive for a fly,” says Marinus L. de Jager of Stellenbosch University in South Africa. Males’ success locating a female bee fly drops in the presence of deceitful daisies, de Jager and Stellenbosch University colleague Allan Ellis say January 29 in the Proceedings of the Royal Society B. That’s the first clear demonstration of sexual deceit’s cost to a pollinator, Ellis says. Such evolutionary costs might push the bee fly to learn from mating mistakes. How long bee flies stay daisy-wary remains unknown. In other studies, wasps tricked by an Australian orchid forgot their lesson after about 24 hours. © Society for Science & the Public 2000 - 2014
Alison Abbott By slicing up and reconstructing the brain of Henry Gustav Molaison, researchers have confirmed predictions about a patient that has already contributed more than most to neuroscience. No big scientific surprises emerge from the anatomical analysis, which was carried out by Jacopo Annese of the Brain Observatory at the University of California, San Diego, and his colleagues, and published today in Nature Communications1. But it has confirmed scientists’ deductions about the parts of the brain involved in learning and memory. “The confirmation is surely important,” says Richard Morris, who studies learning and memory at the University of Edinburgh, UK. “The patient is a classic case, and so the paper will be extensively cited.” Molaison, known in the scientific literature as patient H.M., lost his ability to store new memories in 1953 after surgeon William Scoville removed part of his brain — including a large swathe of the hippocampus — to treat his epilepsy. That provided the first conclusive evidence that the hippocampus is fundamental for memory. H.M. was studied extensively by cognitive neuroscientists during his life. After H.M. died in 2008, Annese set out to discover exactly what Scoville had excised. The surgeon had made sketches during the operation, and brain-imaging studies in the 1990s confirmed that the lesion corresponded to the sketches, although was slightly smaller. But whereas brain imaging is relatively low-resolution, Annese and his colleagues were able to carry out an analysis at the micrometre scale. © 2014 Nature Publishing Group
Henry Molaison, the famous amnesic patient better known as “H.M.,” was unable to form new long-term memories following brain surgery to treat his epilepsy. Scientists who studied his condition made groundbreaking discoveries that revealed how memory works, and before his 2008 death, H.M. and his guardian agreed that his brain would be donated to science. One year after his death, H.M.’s brain was sliced into 2,401 70-micron-thick sections for further study. MIT neuroscience professor emerita Suzanne Corkin studied H.M. during his life and is now part of a team that is analyzing his brain. She is an author of a paper appearing in Nature Communications today reporting preliminary results of the postmortem study. The research team was led by Jacopo Annese at the University of California at San Diego (UCSD). Q: What can we learn from studying H.M.’s brain after his death? And when did you begin laying the groundwork for these postmortem studies? A: It was important to get H.M.’s brain after he died, for three reasons: first of all, to document the exact locus and extent of his lesions, in order to identify the neural substrate for declarative memory. Second, to evaluate the status of the intact brain tissue, revealing the possible brain substrates for the many cognitive functions that H.M. performed normally, including nondeclarative learning without awareness. The third reason was to identify any new abnormalities that occurred as a result of his getting old and were unrelated to the operation. In 1992, I explained to H.M. and his conservator that it would be extremely valuable to have his brain after he died. I told them how important he was to the science of memory, and that he had already made amazing contributions. It would make those even more significant to actually have his brain and see exactly where the damage was. That year, they signed a brain donation form leaving his brain to Massachusetts General Hospital [MGH] and MIT.
Keyword: Learning & Memory
Link ID: 19182 - Posted: 01.29.2014
By BENEDICT CAREY People of a certain age (and we know who we are) don’t spend much leisure time reviewing the research into cognitive performance and aging. The story is grim, for one thing: Memory’s speed and accuracy begin to slip around age 25 and keep on slipping. The story is familiar, too, for anyone who is over 50 and, having finally learned to live fully in the moment, discovers it’s a senior moment. The finding that the brain slows with age is one of the strongest in all of psychology. Lisa Haney Over the years, some scientists have questioned this dotage curve. But these challenges have had an ornery-old-person slant: that the tests were biased toward the young, for example. Or that older people have learned not to care about clearly trivial things, like memory tests. Or that an older mind must organize information differently from one attached to some 22-year-old who records his every Ultimate Frisbee move on Instagram. Now comes a new kind of challenge to the evidence of a cognitive decline, from a decidedly digital quarter: data mining, based on theories of information processing. In a paper published in Topics in Cognitive Science, a team of linguistic researchers from the University of Tübingen in Germany used advanced learning models to search enormous databases of words and phrases. Since educated older people generally know more words than younger people, simply by virtue of having been around longer, the experiment simulates what an older brain has to do to retrieve a word. And when the researchers incorporated that difference into the models, the aging “deficits” largely disappeared. “What shocked me, to be honest, is that for the first half of the time we were doing this project, I totally bought into the idea of age-related cognitive decline in healthy adults,” the lead author, Michael Ramscar, said by email. But the simulations, he added, “fit so well to human data that it slowly forced me to entertain this idea that I didn’t need to invoke decline at all.” © 2014 The New York Times Company
by Helen Thomson The brain that made the greatest contribution to neuroscience and to our understanding of memory has become a gift that keeps on giving. A 3D reconstruction of the brain of Henry Molaison, whose surgery to cure him of epilepsy left him with no short-term memory, will allow scientists to continue to garner insights into the brain for years to come. "Patient HM" became arguably the most famous person in neuroscience after he had several areas of his brain removed in 1953. His resulting amnesia and willingness to be tested have given us unprecedented insights into where memories are formed and stored in the brain. On his death in 2008, HM was revealed to the world as Henry Molaison. Now, a post-mortem examination of his brain, and a new kind of virtual 3D reconstruction, have been published. As a child, Molaison had major epileptic seizures. Anti-epileptic drugs failed, so he sought help from neurosurgeon William Scoville at Hartford Hospital in Connecticut. When Molaison was 27 years old, Scoville removed portions of his medial temporal lobes, which included an area called the hippocampus on both sides of his brain. As a result, Molaison's epilepsy became manageable, but he could not form any new memories, a condition known as anterograde amnesia. He also had difficulty recollecting his long-term past – partial retrograde amnesia.
Keyword: Learning & Memory
Link ID: 19172 - Posted: 01.27.2014
by Bethany Brookshire There are some scientific topics that are bound to generate excitement. A launch to the moon, a potential cure for cancer or any study involving chocolate will always make the news. And then of course there’s caffeine. More than half of Americans have a daily coffee habit, not to mention the boost offered by tea, soda, chocolate and energy drinks. We’d all love to believe that it has more benefit than just papering over a poor night’s sleep. This week, scientists reported that caffeine could give a jolt to memory consolidation, the step right after your brain acquires a memory. During memory consolidation, activity patterns laid down in your brain become more permanent. The study suggested that caffeine might perk up this stage of memory formation. But while it’s an interesting finding, the scientific brew may not be strong enough to justify your coffee habit. Caffeine is a great way to wake you up. It blocks the action of adenosine, a chemical messenger that promotes sleep. Caffeine also has indirect effects on other chemical messengers such as norepinephrine, the neurotransmitter that gives us our famous “fight or flight” response. The net result is increased attention, wakefulness and faster responses. But attention, focus and response time are not memory. And previous studies of memory, says neuroscientist Michael Yassa, the lead author on the new study, were “all over the place.” So Yassa, then at Johns Hopkins University (he’s now at the University of California, Irvine), and undergraduate student Daniel Borota decided to study the effects of caffeine on memory “in a rigorous way.” © Society for Science & the Public 2000 - 2014
A clean slate—that’s what people suffering from posttraumatic stress disorder (PTSD) crave most with their memories. Psychotherapy is more effective at muting more recent traumatic events than those from long ago, but a new study in mice shows that modifying the molecules that attach to our DNA may offer a route to quashing painful memories in both cases. One of the most effective treatments for PTSD is exposure psychotherapy. A behavioral psychologist asks a patient to recall and confront a traumatic event; each time the traumatic memory is revisited, it becomes susceptible to editing through a phenomenon known as memory reconsolidation. As the person relives, for example, a car crash, the details of the event—such as the color and make of the vehicle—gradually uncouple from the anxiety, reducing the likelihood of a panic attack the next time the patient sees, say, a red Mazda. Repeated therapy sessions can also lead to memory extinction, in which the fears tied to an event fade away as old memories are replaced with new ones. Yet this therapy works only for recent memories. If too much time passes before intervention, the haunting visions become stalwart, refusing to budge from the crevices of the mind. This persistence raises the question of how the brain tells the age of a memory in the first place. Researchers at the Massachusetts Institute of Technology, led by neurobiologist Li-Huei Tsai, have now uncovered a chemical modification of DNA that regulates gene activity and dictates whether a memory is too old for reconsolidation in mice. A drug that tweaks these “memory wrinkles” gives old memories a face-lift, allowing them to be edited by reconsolidation and resulting in fear extinction during behavior therapy. © 2014 American Association for the Advancement of Science.
By Ashutosh Jogalekar Popular wisdom holds that caffeine enhances learning, alertness and retention, leading millions to consume coffee or caffeinated drinks before a challenging learning task such as attending a business strategy meeting or a demanding scientific presentation. However a new study in the journal Nature Neuroscience conducted by researchers from Johns Hopkins hints that when it comes to long-term memory and caffeine, timing may be everything; caffeine may enhance consolidation of memories only if it is consumed after a learning or memory challenge. In the study the authors conducted a randomized, double-blind controlled experiment in which 160 healthy female subjects between the ages of 18 and 30 were asked to perform a series of learning tasks. The subjects were handed cards with pictures of various random indoor and outdoor objects (for instance leaves, ducks and handbags) on them and asked to classify the objects as indoor or outdoor. Immediately after the task the volunteers were handed pills, either containing 200 mg of caffeine or placebo. Saliva samples to test for caffeine and its metabolites were collected after 1, 3 and 24 hours. After 24 hours the researchers tested the participants’ recollection of the past day’s test. Along with the items in the test (‘old’) they were presented with new items (‘foils’) and similar looking items (‘lures’), neither of which were part of the task. They were then asked to again classify the items as old, new and similar. There was a statistically significant percentage of volunteers in the caffeinated group that was more likely to mark the ‘similar’ items as ‘similar’ rather than ‘old’. That is, caffeinated participants were clearly able to distinguish much better between the old and the other items, indicating that they were retaining the memory of the old items much better than the people in the placebo group. © 2014 Scientific American,
Training to improve cognitive abilities in older people lasted to some degree 10 years after the training program was completed, according to results of a randomized clinical trial supported by the National Institutes of Health. The findings showed training gains for aspects of cognition involved in the ability to think and learn, but researchers said memory training did not have an effect after 10 years. The report, from the Advanced Cognitive Training for Independent and Vital Elderly (ACTIVE) study, appears in the January 2014 issue of the Journal of the American Geriatrics Society. The project was funded by the National Institute on Aging (NIA) and the National Institute of Nursing Research (NINR), components of the NIH. “Previous data from this clinical trial demonstrated that the effects of the training lasted for five years,” said NIA Director Richard J. Hodes, M.D. “Now, these longer term results indicate that particular types of cognitive training can provide a lasting benefit a decade later. They suggest that we should continue to pursue cognitive training as an intervention that might help maintain the mental abilities of older people so that they may remain independent and in the community.” “ACTIVE is an important example of intervention research aimed at enabling older people to maintain their cognitive abilities as they age,” said NINR Director Patricia Grady, Ph.D. “The average age of the individuals who have been followed over the last 10 years is now 82. Given our nation’s aging population, this type of research is an increasingly high priority.”
Ian Sample, science correspondent A cup or two of coffee could boost the brain's ability to store long-term memories, researchers in the US claim. People who had a shot of caffeine after looking at a series of pictures were better at distinguishing them from similar images in tests the next day, the scientists found. The task gives a measure of how precisely information is stored in the brain, which helps with a process called pattern separation which can be crucial in everyday situations. If the effect is real, and some scientists are doubtful, then it would add memory enhancement to the growing list of benefits that moderate caffeine consumption seems to provide. Michael Yassa, a neuroscientist who led the study at Johns Hopkins University in Baltimore, said the ability to separate patterns was vital for discriminating between similar scenarios and experiences in life. "If you park in the same parking lot every day, the spot you choose can look the same as many others. But when you go and look for your car, you need to look for where you parked it today, not where you parked it yesterday," he said. Writing in the journal Nature Neuroscience, Yassa described how 44 volunteers who were not heavy caffeine consumers and had abstained for at least a day were shown a rapid sequence of pictures on a computer screen. The pictures included a huge range of items, such as a hammer, a chair, an apple, a seahorse, a rubber duck and a car. © 2014 Guardian News and Media Limited