Credit:Sahay Lab, Massachusetts General Hospital, Boston
Play the first few bars of any widely known piece of music, be it The Star-Spangled Banner, Beethoven’s Fifth, or The Rolling Stones’ (I Can’t Get No) Satisfaction, and you’ll find that many folks can’t resist filling in the rest of the melody. That’s because the human brain thrives on completing familiar patterns. But, as we grow older, our pattern completion skills often become more error prone.
This image shows some of the neural wiring that controls pattern completion in the mammalian brain. Specifically, you’re looking at a cross-section of a mouse hippocampus that’s packed with dentate granule neurons and their signal-transmitting arms, called axons, (light green). Note how the axons’ short, finger-like projections, called filopodia (bright green), are interacting with a neuron (red) to form a “memory trace” network. Functioning much like an online search engine, memory traces use bits of incoming information, like the first few notes of a song, to locate and pull up more detailed information, like the complete song, from the brain’s repository of memories in the cerebral cortex.
Credit: Wellcome Centre for Human Neuroimaging, University College London.
In recent years, researchers fueled by the BRAIN Initiative and many other NIH-supported efforts have made remarkable progress in mapping the human brain in all its amazing complexity. Now, a powerful new imaging technology promises to further transform our understanding [1]. This wearable scanner, for the first time, enables researchers to track neural activity in people in real-time as they do ordinary things—be it drinking tea, typing on a keyboard, talking to a friend, or even playing paddle ball.
This new so-called magnetoencephalography (MEG) brain scanner, which looks like a futuristic cross between a helmet and a hockey mask, is equipped with specialized “quantum” sensors. When placed directly on the scalp surface, these new MEG scanners can detect weak magnetic fields generated by electrical activity in the brain. While current brain scanners weigh in at nearly 1,000 pounds and require people to come to a special facility and remain absolutely still, the new system weighs less than 2 pounds and is capable of generating 3D images even when a person is making motions.
Caption: MEMOIR cells variably activate (cyan). The recorded information is then read out to visualize certain RNA transcripts (red). Credit: Elowitz and Cai Labs, Caltech, Pasadena, CA
One of the most fascinating challenges in biology is understanding how a single cell divides and differentiates to form a complex, multicellular organism. Scientists can learn a lot about this process by tracking time-lapse images through a microscope. But gazing through a lens has its limitations, especially in the brain and other opaque and inaccessible tissues and organs.
With support from a 2017 NIH Director’s Transformative Research Program, a California Institute of Technology (Caltech) team now has a way around this problem. Rather than watching or digging information out of cells, the team has learned how to program cells to write their own molecular memoirs. These cells store the information right in their own genomic hard drives. Even better, that information is barcoded, allowing researchers to read it out of the cells without dissecting tissue. The programming can be performed in many different cell types, including stem or adult cells in tissues throughout the body.
Whether it’s lacing up for a morning run, eating blueberry scones, or cheering on the New England Patriots, Steve Ramirez loves life and just about everything in it. As an undergraduate at Boston University, this joie de vivre actually made Ramirez anxious about choosing just one major. A serendipitous conversation helped him realize that all of the amazing man-made stuff in our world has a common source: the human brain.
So, Ramirez decided to pursue neuroscience and began exploring the nature of memory. Employing optogenetics (using light to control brain cells) in mice, he tagged specific neurons that housed fear-inducing memories, making the neurons light sensitive and amenable to being switched on at will.
In groundbreaking studies that earned him a spot in Forbes 2015 “30 Under 30” list, Ramirez showed that it’s possible to reactivate memories experimentally in a new context, recasting them in either a more negative or positive behavior-changing light [1–3]. Now, with support from a 2016 NIH Director’s Early Independence Award, Ramirez, who runs his own lab at Boston University, will explore whether activating good memories holds promise for alleviating chronic stress and psychiatric disease.
Credit: Raunak Basu, University of Utah, Salt Lake City
The final frontier? Trekkies would probably say it’s space, but mapping the brain—the most complicated biological structure in the known universe—is turning out to be an amazing adventure in its own right. Not only are researchers getting better at charting the brain’s densely packed and varied cellular topography, they are starting to identify the molecules that neurons use to connect into the distinct information-processing circuits that allow all walks of life to think and experience the world.
This image shows distinct neural connections in a cross section of a mouse’s hippocampus, a region of the brain involved in the memory of facts and events. The large, crescent-shaped area in green is hippocampal zone CA1. Its highly specialized neurons, called place cells, serve as the brain’s GPS system to track location. It appears green because these neurons express cadherin-10. This protein serves as a kind of molecular glue that likely imparts specific functional properties to this region. [1]
Caption: Colorized 3D reconstruction of dendrites. Neurons receive input from other neurons through synapses, most of which are located along the dendrites on tiny projections called spines. Credit: The Center for Sleep and Consciousness, University of Wisconsin-Madison School of Medicine
People spend about a third of their lives asleep. When we get too little shut-eye, it takes a toll on attention, learning and memory, not to mention our physical health. Virtually all animals with complex brains seem to have this same need for sleep. But exactly what is it about sleep that’s so essential?
Two NIH-funded studies in mice now offer a possible answer. The two research teams used entirely different approaches to reach the same conclusion: the brain’s neural connections grow stronger during waking hours, but scale back during snooze time. This sleep-related phenomenon apparently keeps neural circuits from overloading, ensuring that mice (and, quite likely humans) awaken with brains that are refreshed and ready to tackle new challenges.
It’s not every day that an amateur guitar picker gets to play a duet with an internationally renowned classical cellist. But that was my thrill this week as I joined Yo-Yo Ma in a creative interpretation of the traditional song, “How Can I Keep from Singing?” Our short jam session capped off Mr. Ma’s appearance as this year’s J. Edward Rall Cultural Lecture.
The event, which counts The Dalai Lama, Maya Angelou, and Atul Gawande among its distinguished alumni, this year took the form of a conversation on the intersection of music and science—and earned a standing ovation from a packed house of researchers, patients, and staff here on the National Institutes of Health (NIH) campus in Bethesda, MD.
Caption: Networks of neurons in the mouse retina. Green cells form a special electrically coupled network; red cells express a distinctive fluorescent marker to distinguish them from other cells; blue cells are tagged with an antibody against an enzyme that makes nitric oxide, important in retinal signaling. Such images help to identify retinal cell types, their signaling molecules, and their patterns of connectivity. Credit: Jason Jacoby and Gregory Schwartz, Northwestern University
For Gregory Schwartz, working in total darkness has its benefits. Only in the pitch black can Schwartz isolate resting neurons from the eye’s retina and stimulate them with their natural input—light—to get them to fire electrical signals. Such signals not only provide a readout of the intrinsic properties of each neuron, but information that enables the vision researcher to deduce how it functions and forges connections with other neurons.
The retina is the light-sensitive neural tissue that lines the back of the eye. Although only about the size of a postage stamp, each of our retinas contains an estimated 130 million cells and more than 100 distinct cell types. These cells are organized into multiple information-processing layers that work together to absorb light and translate it into electrical signals that stream via the optic nerve to the appropriate visual center in the brain. Like other parts of the eye, the retina can break down, and retinal diseases, including age-related macular degeneration, retinitis pigmentosa, and diabetic retinopathy, continue to be leading causes of vision loss and blindness worldwide.
In his lab at Northwestern University’s Feinberg School of Medicine, Chicago, Schwartz performs basic research that is part of a much larger effort among vision researchers to assemble a parts list that accounts for all of the cell types needed to make a retina. Once Schwartz and others get closer to wrapping up this list, the next step will be to work out the details of the internal wiring of the retina to understand better how it generates visual signals. It’s the kind of information that holds the key for detecting retinal diseases earlier and more precisely, fixing miswired circuits that affect vision, and perhaps even one day creating an improved prosthetic retina.
Caption: Map of 180 areas in the left and right hemispheres of the cerebral cortex. Credit: Matthew F. Glasser, David C. Van Essen, Washington University Medical School, Saint Louis, Missouri
Neuroscientists have been working for a long time to figure out how the human brain works, and that has led many through the years to attempt to map its various regions and create a detailed atlas of their complex geography and functions. While great progress has been made in recent years, existing brain maps have remained relatively blurry and incomplete, reflecting only limited aspects of brain structure or function and typically in just a few people.
In a study reported recently in the journal Nature, an NIH-funded team of researchers has begun to bring this map of the human brain into much sharper focus [1]. By combining multiple types of cutting-edge brain imaging data from more than 200 healthy young men and women, the researchers were able to subdivide the cerebral cortex, the brain’s outer layer, into 180 specific areas in each hemisphere. Remarkably, almost 100 of those areas had never before been described. This new high-resolution brain map will advance fundamental understanding of the human brain and will help to bring greater precision to the diagnosis and treatment of many brain disorders.
Caption: PET scan images show distribution of tau (top panel) and beta-amyloid (bottom panel) across a brain with early Alzheimer’s disease. Red indicates highest levels of protein binding, dark blue the lowest, yellows and oranges indicate moderate binding. Credit: Brier et al., Sci Transl Med
In people with Alzheimer’s disease, changes in the brain begin many years before the first sign of memory problems. Those changes include the gradual accumulation of beta-amyloid peptides and tau proteins, which form plaques and tangles that are considered hallmarks of the disease. While amyloid plaques have received much attention as an early indicator of disease, until very recently there hadn’t been any way during life to measure the buildup of tau protein in the brain. As a result, much less is known about the timing and distribution of tau tangles and its relationship to memory loss.
Now, in a study published in Science Translational Medicine, an NIH-supported research team has produced some of the first maps showing where tau proteins build up in the brains of people with early Alzheimer’s disease [1]. The new findings suggest that while beta-amyloid remains a reliable early sign of Alzheimer’s disease, tau may be a more informative predictor of a person’s cognitive decline and potential response to treatment.