Star that exploded in 1437 tracked to its current position

Some stars erupt like clockwork. Astronomers have tracked down a star that Korean astronomers saw explode nearly 600 years ago and confirmed that it has had more outbursts since. The finding suggests that what were thought to be three different stellar objects actually came from the same object at different times, offering new clues to the life cycles of stars.

On March 11, 1437, Korean royal astronomers saw a new “guest star” in the tail of the constellation Scorpius. The star glowed for 14 days, then faded. The event was what’s known as a classical nova explosion, which occurs when a dense stellar corpse called a white dwarf steals enough material from an ordinary companion star for its gas to spontaneously ignite. The resulting explosion can be up to a million times as bright as the sun, but unlike supernovas, classical novas don’t destroy the star.
Astronomer Michael Shara of the American Museum of Natural History in New York City and colleagues used digitized photographic plates dating from as early as 1923 to trace a modern star back to the nova. The team tracked a single star as it moved away from the center of a shell of hot gas, the remnants of an old explosion, thus showing that the star was responsible for the nova. The researchers also saw the star, which they named Nova Scorpii AD 1437, give smaller outbursts called dwarf novas in the 1930s and 1940s. The findings were reported in the Aug. 31 Nature.

The discovery fits with a proposal Shara and colleagues made in the 1980s. They suggested that three different stellar observations — bright classical nova explosions, dwarf nova outbursts and an intermediate stage where a white dwarf is not stealing enough material to erupt — are all different views of the same system.

“In biology, we might say that an egg, a larva, a pupa and a butterfly are all the same system seen at different stages of development,” Shara says.

Learning takes brain acrobatics

Peer inside the brain of someone learning. You might be lucky enough to spy a synapse pop into existence. That physical bridge between two nerve cells seals new knowledge into the brain. As new information arrives, synapses form and strengthen, while others weaken, making way for new connections.

You might see more subtle changes, too, like fluctuations in the levels of signaling molecules, or even slight boosts in nerve cell activity. Over the last few decades, scientists have zoomed in on these microscopic changes that happen as the brain learns. And while that detailed scrutiny has revealed a lot about the synapses that wire our brains, it isn’t enough. Neuroscientists still lack a complete picture of how the brain learns.

They may have been looking too closely. When it comes to the neuroscience of learning, zeroing in on synapse action misses the forest for the trees.

A new, zoomed-out approach attempts to make sense of the large-scale changes that enable learning. By studying the shifting interactions between many different brain regions over time, scientists are beginning to grasp how the brain takes in new information and holds onto it.
These kinds of studies rely on powerful math. Brain scientists are co-opting approaches developed in other network-based sciences, borrowing tools that reveal in precise, numerical terms the shape and function of the neural pathways that shift as human brains learn.

“When you’re learning, it doesn’t just require a change in activity in a single region,” says Danielle Bassett, a network neuroscientist at the University of Pennsylvania. “It really requires many different regions to be involved.” Her holistic approach asks, “what’s actually happening in your brain while you’re learning?” Bassett is charging ahead to both define this new field of “network neuroscience” and push its boundaries.

“This line of work is very promising,” says neuroscientist Olaf Sporns of Indiana University Bloomington. Bassett’s research, he says, has great potential to bridge gaps between brain-imaging studies and scientists’ understanding of how learning happens. “I think she’s very much on the right track.”
Already, Bassett and others have found tantalizing hints that the brains that learn best have networks that are flexible, able to rejigger connections on the fly to allow new knowledge in. Some brain regions always communicate with the same neural partners, rarely switching to others. But brain regions that exhibit the most flexibility quickly swap who they’re talking with, like a parent who sends a birthday party invite to the preschool e-mail list, then moments later, shoots off a work memo to colleagues.

In a few studies, researchers have witnessed this flexibility in action, watching networks reconfigure as people learn something while inside a brain scanner. Network flexibility may help several types of learning, though too much flexibility may be linked to disorders such as schizophrenia, studies suggest.

Not surprisingly, some researchers are rushing to apply this new information, testing ways to boost brain flexibility for those of us who may be too rigid in our neural connections.

“These are pretty new ideas,” says cognitive neuroscientist Raphael Gerraty of Columbia University. The mathematical and computational tools required for this type of research didn’t exist until recently, he says. So people just weren’t thinking about learning from a large-scale network perspective. “In some ways, it was a pretty boring mathematical, computational roadblock,” Gerraty says. But now the road is clear, opening “this conceptual avenue … that people can now explore.”
It takes a neural village
That conceptual avenue is more of a map, made of countless neural roads. Even when a person learns something very simple, large swaths of the brain jump in to help. Learning an easy sequence of movements, like tapping out a brief tune on a keyboard, prompts activity in the part of the brain that directs finger movements. The action also calls in brain areas involved in vision, decision making, memory and planning. And finger taps are a pretty basic type of learning. In many situations, learning calls up even more brain areas, integrating information from multiple sources, Gerraty says.

He and colleagues caught glimpses of some of these interactions by scanning the brains of people who had learned associations between two faces. Only one of the faces was then paired with a reward. In later experiments, the researchers tested whether people could figure out that the halo of good fortune associated with the one face also extended to the face it had been partnered with earlier. This process, called “transfer of learning,” is something that people do all the time in daily life, such as when you’re wary of the salad at a restaurant that recently served tainted cheese.

Study participants who were good at applying knowledge about one thing — in this case, a face — to a separate thing showed particular brain signatures, Gerraty and colleagues reported in 2014 in the Journal of Neuroscience. Connections between the hippocampus, a brain structure important for memory, and the ventromedial prefrontal cortex, involved in self-control and decision making, were weaker in good learners than in people who struggled to learn. The scans, performed several days after the learning task, revealed inherent differences between brains, the researchers say. The experiment also turned up other neural network differences among these regions and larger-scale networks that span the brain.

Children who have difficulty learning math, when scanned, also show unexpected brain connectivity, according to research by neuroscientist Vinod Menon of Stanford University and colleagues. Compared with kids without disabilities, children with developmental dyscalculia who were scanned while doing math problems had more connections, particularly among regions involved in solving math problems. That overconnectivity, described in 2015 in Developmental Science, was a surprise, Menon says, since earlier work had suggested that these math-related networks were too weak. But it may be that too many links create a system that can’t accommodate new information. “The idea is that if you have a hyperconnected system, it’s not going to be as responsive,” he says.
There’s a balance to be struck, Menon says. Neural pathways that are too weak can’t carry necessary information, and pathways that are too connected won’t allow new information to move in. But the problem isn’t as simple as that. “It’s not that everything is changing everywhere,” he says. “There is a specificity to it.” Some connections are more important than others, depending on the task.

Neural networks need to shuttle information around quickly and fluidly. To really get a sense of this movement as opposed to snapshots frozen in time, scientists need to watch the brain as it learns. “The next stage is to figure out how the networks actually shift,” Menon says. “That’s where the studies from Dani Bassett and others will be very useful.”

Flexing in real time
Bassett and colleagues have captured these changing networks as people learn. Volunteers were given simple sequences to tap out on a keyboard while undergoing a functional MRI scan. During six weeks of scanning as people learned the task, neural networks in their brains shifted around. Some connections grew stronger and some grew weaker, Bassett and her team reported in Nature Neuroscience in 2015.

People who quickly learned to tap the correct sequence of keys showed an interesting neural trait: As they learned, they shed certain connections between their frontal cortex, the outermost layer of the brain toward the front of the head, and the cingulate, which sits toward the middle of the brain. This connection has been implicated in directing attention, setting goals and making plans, skills that may be important for the early stages of learning but not for later stages, Bassett and colleagues suspect. Compared with slow learners, fast learners were more likely to have shunted these connections, a process that may have made their brains more efficient.

Flexibility seems to be important for other kinds of learning too. Reinforcement learning, in which right answers get a thumbs up and wrong answers are called out, also taps into brain flexibility, Gerraty, Bassett and others reported online May 30 at bioRxiv.org. This network comprises many points on the cortex, the brain’s outer layer, and a deeper structure known as the striatum. Other work on language comprehension, published by Bassett and colleagues last year in Cerebral Cortex, found some brain regions that were able to quickly form and break connections.

These studies captured brains in the process of learning, revealing “a much more interesting network structure than what we previously thought when we were only looking at static snapshots,” Gerraty says. The learning brain is incredibly dynamic, he says, with modules breaking off from partners and finding new ones.

While the details of those dynamics differ from study to study, there is an underlying commonality: “It seems that part of learning about the world is having parts of your brain become more flexible, and more able to communicate with different areas,” Gerraty says. In other words, the act of learning takes flexibility.

But too much of a good thing may be bad. While performing a recall task in a scanner, people with schizophrenia had higher flexibility among neural networks across the brain than did healthy people, Bassett and colleagues reported last year in the Proceedings of the National Academy of Sciences. “That suggests to me that while flexibility is good for healthy people, there is perhaps such a thing as too much flexibility,” Bassett says.
Just how this flexibility arises, and what controls it, is unknown. Andrea Stocco, a cognitive neuroscientist at the University of Washington in Seattle, suspects that a group of brain structures called the basal ganglia, deep within the brain, has an important role in controlling flexibility. He compares this region, which includes the striatum, to an air traffic controller who shunts information to where it’s most needed. One of the basal ganglia’s jobs seems to be shutting things down. “Most of the time, the basal ganglia is blocking something,” he says. Other researchers have found evidence that crucial “hubs” in the cortex help control flexibility.

Push for more
Researchers don’t yet know how measures of flexibility in brain regions relate to the microscopic changes that accompany learning. For now, the macro and the micro views of learning are separate worlds. Despite that missing middle ground, researchers are charging ahead, looking for signs that neural flexibility might offer a way to boost learning aptitude.

It’s possible that external brain stimulation may enhance flexibility. After receiving brain stimulation carefully aimed at a known memory circuit, people were better able to recall lists of words, scientists reported May 8 in Current Biology. If stimulation can boost memory, some argue, the technique could enhance flexibility and perhaps learning too.
Certain drugs show promise. DXM, found in some cough medicines, blocks proteins that help regulate nerve cell chatter. Compared with a placebo, the compound made some brain regions more flexible and able to rapidly switch partners in healthy people, Bassett and colleagues reported last year in the Proceedings of the National Academy of Sciences. She is also studying whether neurofeedback — a process in which people try to change their brain patterns to become more flexible with real-time monitoring — can help.

Something even simpler might work for boosting flexibility. On March 31 in Scientific Reports, Bassett and colleagues described their network analyses of an unusual subject. For a project called MyConnectome, neuroscientist Russ Poldrack, then at the University of Texas at Austin, had three brain scans a week for a year while assiduously tracking measures that included mood. Bassett and her team applied their mathematical tools to Poldrack’s data to get measurements of his neural flexibility on any given scan day. The team then looked for associations with mood. The standout result: When Poldrack was happiest, his brain was most flexible, for reasons that aren’t yet clear. (Flexibility was lowest when he was surprised.)

Those results are from a single person, so it’s unknown how well they would generalize to others. What’s more, the study identifies only a link, not that happiness causes more flexibility or vice versa. But the idea is intriguing, if not obvious, Bassett says. “Of course, no teacher is really going to say we’re doing rocket science if we tell them we should make the kids happier and then they’ll learn better.” But finding out exactly how happiness relates to learning is important, she says.

The research is just getting started. But already, insights on learning are coming quickly from the small group of researchers viewing the brain as a matrix of nodes and links that deftly shift, swap and rearrange themselves. Zoomed out, network science brings to the brain “a whole new set of hypotheses and new ways of testing them,” Bassett says.

Microbes hobble a widely used chemo drug

Some bacteria may shield tumor cells against a common chemotherapy drug.

Certain types of bacteria make an enzyme that inactivates the drug gemcitabine, researchers report in the Sept. 15 Science. Gemcitabine is used to treat patients with pancreatic, lung, breast and bladder cancers.

Bacteria that produce the enzyme cytidine deaminase converted the drug to an inactive form. That allowed tumor cells to survive gemcitabine treatment in lab dishes and mouse studies, Leore Geller of the Weizmann Institute of Science in Rehovot, Israel, and colleagues discovered. More than 98 percent of the enzyme-producing microbes belong to the Gammaproteobacteria class, which includes E. coli and about 250 bacterial genera.
Pancreatic tumors taken from human patients also carried the enzyme-producing bacteria. Of 113 pancreatic ductal adenocarcinoma samples studied, 86 contained gemcitabine-inactivating bacteria.

Antibiotics may correct the problem. In the study, Geller and colleagues infected mice that had colon cancer with the enzyme-producing bacteria. Tumors grew rapidly in infected mice treated with gemcitabine alone. Giving the mice antibiotics helped gemcitabine kill tumor cells, increasing the number of tumor cells going through a type of cell death called apoptosis from about 15 percent to 60 percent or more. That result may indicate that combinations of gemcitabine and antibiotics could make chemotherapy more effective for some cancer patients.

Confusion lingers over health-related pros and cons of marijuana

No one knows whether chronic marijuana smoking causes emotional troubles or is a symptom of them…. This dearth of evidence has a number of explanations: serious lingering reactions, if they exist, occur after prolonged use, rarely after a single dose; marijuana has no known medical use, unlike LSD, so scientists have had little reason to study the drug…. Also, marijuana has been under strict legal sanctions … for more than 30 years. – Science News, October 7, 1967

In 29 states and in Washington, D.C., marijuana is now commonly prescribed for post-traumatic stress disorder and chronic pain. But the drug’s pros and cons remain hazy. Regular pot use has been linked to psychotic disorders and to alcohol and drug addiction (SN Online: 1/12/17). And two recent research reviews conclude that very little high-quality data exist on whether marijuana effectively treats PTSD or pain. Several large-scale trials are under way to assess how well cannabis treats these conditions.

Body clock mechanics wins U.S. trio the Nobel Prize in physiology or medicine

Discoveries about the clocklike ups and downs of daily life have won Jeffery C. Hall, Michael Rosbash and Michael W. Young the Nobel Prize in physiology or medicine.

Circadian rhythms are daily cycles of hormones, gene activity and other biological processes that govern sleep, body temperature and metabolism. When thrown out of whack, there can be serious health consequences, including increased risk of diabetes, heart and Alzheimer’s diseases.

Hall and Rosbash discovered the first molecular gear of the circadian clockworks: A protein called Period increases and decreases in abundance on a regular cycle during the day. Young discovered that another protein called Timeless works with Period to drive the clock. Young also discovered other circadian clockworks.

Nanoscale glitches let flowers make a blue blur that bees can see

A bit of imperfection could be perfect for flowers creating a “blue halo” effect that bees can see.

At least a dozen families of flowering plants, from hibiscuses to daisy relatives, have a species or more that can create a bluish-ultraviolet tinge using arrays of nanoscale ridges on petals, an international research team reports online October 18 in Nature. These arrays could be the first shown to benefit from the sloppiness of natural fabrication, says coauthor Silvia Vignolini, a physicist specializing in nanoscale optics at the University of Cambridge.
Flowers, of course, can’t reach industrial standards for uniform nanoscale fabrication. Yet the halo may be a case where natural imperfections may be important to a flower’s display. Tests with artificial flowers showed that the nanoglitches made it easier for bees to learn that a showy petal meant a sugary reward, Vignolini and colleagues found.
Blues are rare in actual pigments in living things( SN: 12/10/16, p. 4 ). Color in the wings of Morpho butterflies or blue jay feathers, for instance, comes from nanoscale structures that contain no pigments but create colorful illusions by muting some wavelengths of light while intensely reflecting others ( SN: 6/11/16, p. 32 ).
Flower petals make their blue halo illusion with somewhat irregular versions of what are called diffraction gratings, rows of ridges like the recording surface on a CD. A perfectly regular array of ridges would create true iridescence, changing color depending on the angle a viewer takes. The flowers’ imperfections, variations in ridge height and spacing, weaken or destroy the iridescence. A viewer swooping by would see less color shifting and more of a bluish-ultraviolet tinge reflected at a wider range of angles.

To see whether bees respond more to iridescence or a blue halo, researchers created sets of artificial flowers, pieces of epoxy resin with some kind of nanoscale-ridged array. A petal-scale project was huge compared with the usual nanoscale experiments, requiring marathon fabrication sessions. “We were a pain to everybody,” Vignolini says.

In two tests, researchers offered bumblebees a pair of “flowers,” one that held sugar water and one with a nasty-tasting solution, to see how quickly bees would learn to distinguish sweet from foul. When the flower’s nanoridges had imperfections creating a blue halo, bees learned the task faster than when the flower had perfect iridescence. Imperfect arrays were actually an advantage for the flowers in creating displays pollinating bees find memorable, the researchers conclude.
Such disorder in nature’s structural color (versus pigments) has shown up before, as in obviously jumbled color-trick structures in bird feathers. Before the tests, though, it was unclear whether flowers would benefit from perfect iridescence and were just falling short in growing perfect arrays. The blue halo might have been merely a side effect of challenging botanical fabrication. The bee experiments, however, showed the opposite, the researchers say. These are the first tests to show that some disorder is not just a downside of natural fabrication but in itself “has a function,” Vignolini says.

That result makes sense to visual ecologist Nathan Morehouse of the University of Cincinnati. Nanostructures that iridesce may often just be a way birds or butterflies can create an unusual color rather than a way to produce iridescence for its own sake. The shifting colors might even have a downside. By definition, true iridescence changes color as an insect or bird changes its angle of approach, and so may not be the best form for an easy-to-remember signal. “Iridescence itself is something they just have to manage,” he suggests.

Alligators eat sharks — and a whole lot more

Alligators don’t just stick to freshwater and the prey they find there. These crafty reptiles can live quite easily, at least for a bit, in salty waters and find plenty to eat — including crabs, sea turtles and even sharks.

“They should change the textbooks,” says James Nifong, an ecologist with the Kansas Cooperative Fish and Wildlife Research Unit at Kansas State University in Manhattan, who has spent years documenting the estuarine gator diet.

Nifong’s most recent discovery, splashed all over the news last month, is that the American alligator (Alligator mississippiensis) eats at least three species of shark and two species of rays, he and wildlife biologist Russell Lowers report in the September Southeastern Naturalist.

Lowers captured a female gator with a young Atlantic stingray in her jaws near where he works at Kennedy Space Center in Cape Canaveral, Florida. And he and Nifong gathered several other eyewitness accounts: A U.S. Fish and Wildlife employee spotted a gator consuming a nurse shark in a Florida mangrove swamp in 2003. A birder photographed an alligator eating a bonnethead shark in a Florida salt marsh in 2006. One of Nifong’s collaborators, a marine turtle researcher, saw gators consuming both bonnethead and lemon sharks in the late 1990s. And Nifong found yet another report of a gator eating a bonnethead shark in Hilton Head, S.C., after their paper was published. All of these snacks required gators to venture into salty waters.
But shark may not be the most surprising item on the alligator estuarine menu. Nifong spent years catching hundreds of wild gators and pumping their stomachs to figure out what they eat, work that relies “on electrical tape, duct tape and zip ties,” Nifong says. And he found that the menu is pretty long.

To snag an alligator, he uses a big blunted hook or, with smaller animals, just grabs the animal and hauls it into the boat. He gets a noose around its neck. Then the researchers tape the mouth shut, take body measurements (everything from weight to toe length) and get blood or urine samples.

Once that’s out of the way, the team will strap the gator to a board with Velcro ties or rope. Then, it’s time to untape the mouth, quickly insert a piece of pipe to hold it open, and tape the alligator’s mouth around the pipe. The pipe, Nifong says, is there “so they can’t bite down.” And that’s important, because next someone has to stick a tube down the gator’s throat and hold it there to keep the animal’s throat open.
Finally, “we fill [the stomach] up with water very slowly so we don’t injure the animal,” Nifong says. “Then we do basically the Heimlich maneuver.” Pressing down on the abdomen forces the gator to give up its stomach contents. Usually.
“Sometimes it goes better than other times,” he says. “They can just decide to not let it out.” Then the researchers carefully undo all their work to let the gator loose.

Back in the lab, Nifong and his colleagues teased out what they could find in those stomach contents, and looked for more clues about the animals’ diet from in the blood samples. Nifong and his colleagues found that the gators were eating a rich marine diet, including small fish, mammals, birds, insects and crustaceans. They’ll even eat fruit and seeds. The sharks and rays didn’t show up in these studies (nor did sea turtles, which gators have also been spotted munching on). But Nifong and Lowers speculate that’s because the tissue of those animals gets digested very quickly. So if a gator had eaten a shark more than a few days before being caught, there was no way to know.

Because alligators don’t have any salt glands, “they’re subject to the same pressures as me or you when being out in saltwater,” Nifong says. “You’re losing water, and you’re increasing salt in your blood system.” That can lead to stress and even death, he notes. So the gators tend to just go back and forth between saltwater and freshwater. They can also close off their throat with a cartilaginous shield and shut their nostrils to keep salty water out. And when they eat, they’ll tip their head up to let the saltwater drain out before gulping down their catch.
What alligators eat isn’t as important a finding as the discovery that they regularly travel between saltwater and freshwater environments, Nifong says. And, he notes, “it occurs across a wide variety of habitats across the U.S. southeast.” That’s important because the gators are moving nutrients from rich marine waters into poorer, fresh waters. And they may be having a larger effect on estuarine food webs that anyone had imagined.

For instance, one of the prey items on the alligator menu is blue crab. Gators “scare the bejesus out of them,” Nifong says. And when gators are around, blue crabs decrease their predation of snails, which might then eat more of the cordgrass that forms the base of the local ecosystem. “Understanding that an alligator has a role in that kind of interaction,” Nifong points out, is important when planning conservation efforts.

Artificial insulin-releasing cells may make it easier to manage diabetes

Artificial cells made from scratch in the lab could one day offer a more effective, patient-friendly diabetes treatment.

Diabetes, which affects more than 400 million people around the world, is characterized by the loss or dysfunction of insulin-making beta cells in the pancreas. For the first time researchers have created synthetic cells that mimic how natural beta cells sense blood sugar concentration and secrete just the right amount of insulin. Experiments with mice show that these cells can regulate blood sugar for up to five days, researchers report online October 30 in Nature Chemical Biology.
If the mouse results translate to humans, diabetics could inject these artificial beta cells to automatically regulate their blood sugar levels for days at a time.

That would be a “a huge leap forward” for diabetic patients who currently have to check their blood sugar and inject insulin several times a day, says Omid Veiseh, a bioengineer at Rice University in Houston who wasn’t involved in the research. “Even if it were just a one-day thing, it would still be impressive,” he says.
Fashioned from human-made materials and biological ingredients like proteins, these faux cells contain insulin-filled pouches much like the insulin-carrying compartments inside real beta cells. And, similar to a natural beta cell, when one of these artificial beta cells is surrounded by excess blood sugar, its insulin sacs fuse with its outer membrane and eject insulin into the bloodstream. As blood sugar levels drop, insulin packets stop fusing with the membrane, which stems the cell’s insulin secretion.
Fabricating artificial insulin delivery systems that actually imitate the inner workings of real beta cells for ultrafine blood sugar regulation is “an engineering feat,” says Patrik Rorsman, a diabetes researcher at the University of Oxford who wasn’t involved in the work. The cellular imitations are “not as perfect as the beta cells we’re equipped with when we’re healthy,” he adds. For one thing, the faux cells eventually run out of insulin to release. But Rorsman believes that such artificial cells present a viable diabetes treatment.
Unlike transplanted beta cells — or other types of real cells genetically engineered to release insulin for diabetes treatment (SN: 1/15/11, p. 9) — these artificial cells could be mass-produced and have a much longer shelf life than live cells, says study coauthor Zhen Gu, a biomedical engineer at the University of North Carolina at Chapel Hill.

When Gu and colleagues injected their synthetic cells into diabetic mice, the animals’ blood sugar levels normalized within an hour and stayed that way up to five days, when the cells ran out of insulin. The researchers plan to perform further tests on lab animals to assess the fake cells’ long-term health effects before running clinical trials.

Even for patients who manage their insulin with automated mechanical pumps (SN Online: 5/8/10), synthetic cells offer the advantage of more precise, real time blood sugar regulation, says Michael Strano, a bioengineer at MIT. The creation of the new faux cells not only poses a potential diabetes treatment, “but it’s also a bellwether. It’s slightly ahead of its time,” says Strano. “I think therapeutics of the future are going to look like this.”

The Arecibo Observatory will remain open, NSF says

The iconic Arecibo Observatory has survived a hurricane and dodged deep budget cuts. On November 16, the National Science Foundation, which funds the bulk of the observatory’s operating costs, announced that they would continue funding the radio telescope at a reduced level.

It’s not clear yet who will manage the observatory in the long run, or where the rest of the funding will come from. But scientists are celebrating. For example:
Arecibo, a 305-meter-wide radio telescope located about 95 kilometers west of San Juan, is the second largest radio telescope in the world. It has been instrumental in tasks as diverse as monitoring near-Earth asteroids, watching for bright blasts of energy called fast radio bursts and searching for extraterrestrial intelligence.

But the NSF, which covers $8.3 million of the observatory’s nearly $12 million annual budget, has been trying to back away from that responsibility for several years. After Hurricane Maria hit Puerto Rico on September 20, damaging the telescope’s main antenna, the observatory’s future seemed unclear (SN: 9/29/17).

On November 16, the NSF released a statement announcing it would continue science operations at Arecibo “with reduced agency funding,” and would search for new collaborators to cover the rest.、
“This plan will allow important research to continue while accommodating the agency’s budgetary constraints and its core mission to support cutting-edge science and education,” the statement says.

A new map exhibit documents evolving views of Earth’s interior

Much of what happens on the Earth’s surface is connected to activity far below. “Beneath Our Feet,” a temporary exhibit at the Norman B. Leventhal Map Center in the Boston Public Library, explores the ways people have envisioned, explored and exploited what lies underground.

“We’re trying to visualize those places that humans don’t naturally go to,” says associate curator Stephanie Cyr. “Everybody gets to see what’s in the sky, but not everyone gets to see what’s underneath.”
“Beneath Our Feet” displays 70 maps, drawings and archaeological artifacts in a bright, narrow exhibit space. (In total, the library holds a collection of 200,000 maps and 5,000 atlases.) Many objects have two sets of labels: one for adults and one for kids, who are guided by a cartoon rat mascot called Digger Burrows.

The layout puts the planet’s long history front and center. Visitors enter by walking over a U.S. Geological Survey map of North America that is color-coded to show how topography has changed over geologic time.
Beyond that, the exhibit is split into two main themes, Cyr says: the natural world, and how people have put their fingerprints on it. Historical and modern maps hang side by side, illustrating how ways of thinking about the Earth developed as the tools for exploring it improved.

For instance, a 1665 illustration drawn by Jesuit scholar Athanasius Kircher depicts Earth’s water systems as an underground network that churned with guidance from a large ball of fire in the planet’s center, Cyr says. “He wasn’t that far off.” Under Kircher’s drawing is an early sonar map of the seafloor in the Pacific Ocean, made by geologists Marie Tharp and Bruce Heezen in 1969 (SN: 10/6/12, p. 30). Their maps revealed the Mid-Atlantic Ridge. Finding that rift helped to prove the existence of plate tectonics and that Earth’s surface is shaped by the motion of vast subsurface forces.

On another wall, a 1794 topological-relief drawing of Mount Vesuvius — which erupted and destroyed the Roman city of Pompeii in A.D. 79 — is embellished by a cartouche of Greek mythological characters, including one representing death. The drawing hangs above a NASA satellite image of the same region, showing how the cities around Mount Vesuvius have grown since the eruption that buried Pompeii, and how volcano monitoring has improved.

The tone turns serious in the latter half of the exhibit. Maps of coal deposits in 1880s Pennsylvania sit near modern schematics explaining how fracking works (SN: 9/8/12, p. 20). Reproductions of maps of the Dakotas from 1886 may remind visitors of ongoing controversies with the Dakota Access Pipeline, proposed to run near the Standing Rock Sioux Reservation, and maps from the U.S. Environmental Protection Agency mark sites in Flint, Mich., with lead-tainted water.

Maps in the exhibit are presented dispassionately and without overt political commentary. Cyr hopes the zoomed-out perspectives that maps provide will allow people to approach controversial topics with cool heads.

“The library is a safe place to have civil discourse,” she says. “It’s also a place where you have access to factual materials and factual resources.”