When measuring lead in water, check the temperature

Lead contamination in drinking water can change with the seasons. Tracking lead levels in water pipes over several months, researchers discovered three times as much dissolved lead and six times as much undissolved lead in summer than in winter. The finding could help improve water testing, says study coauthor Sheldon Masters, an environmental engineer at Virginia Tech and Corona Environmental Consulting in Philadelphia.

Masters and colleagues analyzed water contamination data collected from pipes in Washington, D.C., and Providence, R.I., and tested the dissolvability of lead in different water conditions. In many, but not all, homes and lab tests, the amount of lead leaching into drinking water rose as water temperature increased.

For pipes in Washington, average wintertime dissolved lead levels were 3.6 parts per billion, compared with 10.8 ppb during summer. Average undissolved lead concentrations varied from 7.6 ppb during winter to 48.4 ppb during summer. Each 1 degree Celsius rise in water temperature boosted dissolved lead levels by about 17 percent and lead particles by about 36 percent, the researchers report online April 14 in Environmental Science & Technology. Washington water temperature varied from about 5° to 30° C. Seasonal variations in lead were smaller than those expected from temperature changes alone, since other factors such as the amount of organic matter in water can also influence lead levels.

Some water systems could meet the regulatory standard of less than 15 ppb in winter while exceeding that threshold during warmer months, the researchers warn. Water testing prioritizes conditions with the highest risk for lead leaching. However, no current guidelines explicitly address seasonal variability. Lead consumption can cause severe health problems including birth defects, anemia and brain damage (SN: 3/19/16, p. 8).

New analysis: Genetically engineered foods not a health risk

Genetically engineered crops don’t appear to harm humans or the environment, according to a new report released May 17 by the National Academies of Sciences, Engineering and Medicine.

An extensive analysis of two decades’ worth of evidence dug up no substantial proof that genetically engineered foods were any less safe to eat than those that are conventionally bred. The study’s authors also found no conclusive causal link between the engineered crops and environmental problems. The authors note, though, that it’s not always easy to make definitive conclusions; measuring long-term environmental changes is complicated.

The news comes in the midst of political tumult in the United States over laws to label foods made with GE ingredients. But when it comes to food safety and the environment, the authors conclude, how a plant is made isn’t as important as what is actually created.

“It is the product, not the process, that should be regulated,” the authors write.

3-D TVs are a work in need of progress

3-D Home TV Foreseen — The pace of new developments in the recently revived method of photography known as holography is so fast that three-dimensional television sets portraying life-size scenes could be a reality before 1984, as was predicted in George Orwell’s novel…. A hologram is a recording of an interference pattern reflected from an object. From this recording, the object can be reconstructed visually in a three-dimensional form. — Science News, June 11, 1966

UPDATE
Television viewers are still waiting for the 3-D revolution. Although 3-D TVs went on sale in the United States and elsewhere in 2010, they have yet to take off. Most sets require special glasses or have limited viewing angles, and none use holography to create the illusion of depth. Scientists haven’t given up, though. Using innovative plastic screens, researchers are projecting small holographic movies in real time (SN: 12/17/11, p. 18). The enormous bandwidth and processing power needed to transmit and display the images are still huge barriers to making Orwell’s vision a reality.

Spy satellites reveal early start to Antarctic ice shelf collapse

The biggest ice shelf collapse on record was set in motion years earlier than previously thought, new research reveals.

Analyzing declassified images from spy satellites, researchers discovered that the downhill flow of ice on Antarctica’s Larsen B ice shelf was already accelerating as early as the 1960s and ’70s. By the late 1980s, the average ice velocity at the front of the shelf was around 20 percent faster than in the preceding decades, the researchers report in a paper to be published in Geophysical Research Letters.
Rising temperatures since the 1950s probably quickened the ice flow, which in turn put more strain on the ice and further weakened the shelf, says study coauthor Hongxing Liu, a geographer at the University of Cincinnati. Previous work had suggested that the ice shelf’s downward slide began only a few years before a Rhode Island-sized region of ice disintegrated into thousands of icebergs in 2002.

The new data will help scientists more confidently predict how Antarctic ice will fare in the coming decades, says Penn State glaciologist Richard Alley, who was not involved in the work. The early response of Larsen B to warming “is consistent with this ice shelf system being sensitive, and gives a target for future modeling studies to learn how sensitive, and for what reasons,” he says.

Ice shelves such as Larsen B line Antarctica’s coast and slow the flow of the continent’s glaciers and ice sheets into the sea. Rising temperatures are shrinking Antarctica’s ice, with several ice shelves on track to disappear completely within 100 years (SN Online: 3/26/15). Tracking the long-term decline of ice shelves is tricky, though. Scientific satellite images are sparse prior to the 1990s and next to nonexistent before the 1980s.

Liu and colleagues turned to another group that peered at Antarctica, a U.S. intelligence agency called the National Reconnaissance Office. In 1963, the agency photographed the continent as part of an intelligence-gathering mission. While these images were declassified in 1995, the photos were too distorted by effects such as the camera used and Earth’s curvature to use for ice flow measurements.

Making the photographs usable required identifying stationary landmarks for reference, a difficult task on a continent covered with shifting white ice. Comparing the spy photos with later scientific images, Liu and colleagues identified 44 potential landmarks. Then, using the locations as anchor points, the researchers unwarped the images. Along with additional satellite images snapped in 1979 and the 1980s, the modified images allowed the researchers to track Larsen B’s ice flow over time.
The ice on Larsen B’s front flowed at around 400 meters per year on average between 1963 and 1986, calculations using images from those years indicate. From 1986 to 1988, the average was 490 meters per year. That speed boost suggests that the ice flow accelerated between the 1963 to 1986 satellite images. Several glaciers that feed into Larsen B underwent similar accelerations, the researchers found.

Larsen B’s early acceleration hints that the ice shelf was already weakening well before the 1990s, says Ted Scambos, a polar scientist at the National Snow and Ice Data Center in Boulder, Colo., who was not involved in the study. Previous studies suggested that balmy surface temperatures caused Larsen B’s demise by forming meltwater pools on top of the ice shelf that forced open cracks in the ice (SN: 10/18/14, p. 9). The new satellite data suggest that this fracturing was a finishing blow following long-term weakening by forces such as relatively warm seawater eroding the ice shelf’s underside, Scambos says.

Moms’ voices get big reactions in kids’ brains

Any parent trying to hustle a school-bound kid out the door in the morning knows that her child’s skull possesses a strange and powerful form of black magic: It can repel parents’ voices. Important messages like “find your shoes” bounce off the impenetrable fortress and drift unheeded to the floor.

But when this perplexing force field is off, it turns out that mothers’ voices actually have profound effects on kids. Children’s brains practically buzz when they hear their moms’ voices, scientists report in the May 31 Proceedings of the National Academy of Sciences. (Fun and not surprising side note: Babies’ voices get into moms’ brains, too.)

The parts of kids’ brains that handle emotions, face recognition and reward were prodded into action by mothers’ voices, brain scans of 24 children ages 7 to 12 revealed. And words were not required to get this big reaction. In the study, children listened to nonsense words said by either their mother or one of two unfamiliar women. Even when the words were fake, mothers’ voices still prompted lots of neural action.

The study was done in older kids, but children are known to tune into their mothers’ voices early. Really early, in fact. One study found that fetuses’ heart rates change when they hear their moms read a story. For a fetus crammed into a dark, muffled cabin, voices may take on outsized importance.

And voices carry particularly powerful messages throughout childhood. “A tremendous amount of emotional information is conveyed to children through auditory channels,” says University of Wisconsin-Madison child psychologist Seth Pollak. And, he points out, kids are small. “Kids’ faces are down around our knees. And children who are crawling are looking at the ground,” he says. This obvious point means that facial expressions and other visual signals might not pack as much punch as a voice.

Of course, voices other than those belonging to moms are also important. Pollak says that voices of fathers — or any other caregiver who spends lots of time around a child — probably affect children’s brains in a similar way. It’s just that those studies haven’t been done yet.

The results of the latest brain scan study make a lot of sense, says Pollak. Some of the brain regions activated are those involved in feeling good. “A caregiver’s voice is actually rewarding. It activates the systems that make us feel calm,” he says.
And the new study might help explain some of Pollak’s earlier results. He and his colleagues stressed out 68 girls, who happened to be the same ages as those in the brain scanning study, by making them do math and word problems in front of three unsmiling adult strangers — a terrifying prospect for most kids. (And adults.) After their ordeal, the girls either talked to their moms in person, on the phone or by instant messenger.

Compared with the instant messenger typers, the girls who saw their moms in person or talked to them on the phone were more soothed, showing lower levels of stress hormones. That finding, published in 2012 in Evolution and Human Behavior, suggests that to a kid, there’s something especially calming about hearing her own mother’s voice.

And now, by showing the widespread reaction to a mother’s voice, the brain data back that up. “It all kind of hangs together in a way that I think is very intuitive,” Pollak says. In other words, a mother’s voice is powerful, perhaps even strong enough to overcome a force field.

Lidar maps vast network of Cambodia’s hidden cities

Thanks to modern laser technology, Southeast Asia’s Khmer Empire is rising from forest floors for the first time in centuries.

New findings show the remarkable extent to which Khmer people built cities and transformed landscapes from at least the fifth to the 15th century, and perhaps for several hundred years after that, says archaeologist Damian Evans of Cambodia’s Siem Reap Center. Laser mapping in 2015 of about 1,910 square kilometers of largely forested land in northern Cambodia indicates that gridded city streets and extensive canals emerged surprisingly early, by around A.D. 500, Evans reports June 13 in the Journal of Archaeological Science. Researchers have generally assumed that large-scale urban development began later at Greater Angkor, capital of the Khmer Empire from the ninth to 15th centuries (SN: 5/14/16, p. 22).
A helicopter carrying light detection and ranging equipment, lidar for short, flew sorties over seven Khmer sites in the vicinity of Greater Angkor. Lidar’s laser pulses gathered data on the contours of jungle- and vegetation-covered land. Lidar maps revealed city blocks, canals and other remnants of past settlements.
Mysterious ground features previously identified by lidar surveys at Angkor Wat temple in Greater Angkor also turned up at several sites, some located as many as 100 kilometers from Greater Angkor. Those sites include the eighth to ninth century city of Mahendraparvata and a 12th century city, Preah Khan of Kompong Svay. Fields of precisely arranged earthen mounds at these settlements may have been used to collect rainwater, Evans speculates. Earthen embankments forming coiled or spiral patterns might have been gardens or ceremonial spaces.

“It’s humbling to see the lidar data and realize how much was previously missed in ground surveys at Preah Khan,” says archaeologist Mitch Hendrickson of the University of Illinois at Chicago. Hendrickson conducts research at Preah Khan, one of several ancient cities that provided food and other services to Greater Angkor via an extensive road system.

Before the 2015 lidar survey, Mahendraparvata was known “only from inscription texts and a few bits of broken-down masonry,” adds archaeologist Charles Higham of the University of Otago in Dunedin, New Zealand. Mahendraparvata’s laser-traced layout indicates it was an early, small-scale version of Greater Angkor, Higham says.
A military invasion and sacking of Greater Angkor in the 15th century apparently did not result in most of its roughly 750,000 residents abandoning the site, as many investigators have thought. Lidar data from 2015 indicate that Khmer capitals established after Greater Angkor’s defeat, such as Longvek and Oudong, show no signs of dense populations created by mass relocations from the former capital, Evans says.

That suggests that the political state collapsed at Greater Angkor, but hundreds of thousands of rice farmers carried on, Hendrickson says. “Lots of fish and rice were still available,” he says. “Local farmers were more resilient than the state was.”

Coral bleaching event is longest on record

Coral reefs won’t be out of hot water anytime soon. A global bleaching event that began in June 2014 is the longest on record and now covers a larger area than ever before. What’s worse, it shows no signs of ending.

Global warming exacerbated by the latest El Niño is to blame, National Oceanic and Atmospheric Administration scientists reported Monday at the 13th International Coral Reef Symposium in Honolulu. Since 1979, periodic mass bleachings covering hundreds of kilometers have only lasted for “a year or so,” said NOAA Coral Reef Watch Coordinator Mark Eakin. But this one has dragged on for two years, threatening more than 40 percent of reefs globally, and more than 70 percent in the United States.

When corals are stressed by heat, they reject the colorful algae living inside them and turn a ghostly white. Those algae are a major source of food, so reefs can die if conditions don’t improve.

NOAA scientists aren’t sure what will end this episode. It could extend into 2017, and more frequent events are possible in the future, the scientists said. “Climate models suggest that most coral reefs may be seeing bleaching every other year by mid-century,” Eakin added. “How much worse that gets will depend on how we deal with global warming.”

Smartphones may be changing the way we think

Not too long ago, the internet was stationary. Most often, we’d browse the Web from a desktop computer in our living room or office. If we were feeling really adventurous, maybe we’d cart our laptop to a coffee shop. Looking back, those days seem quaint.

Today, the internet moves through our lives with us. We hunt Pokémon as we shuffle down the sidewalk. We text at red lights. We tweet from the bathroom. We sleep with a smartphone within arm’s reach, using the device as both lullaby and alarm clock. Sometimes we put our phones down while we eat, but usually faceup, just in case something important happens.
Our iPhones, Androids and other smartphones have led us to effortlessly adjust our behavior. Portable technology has overhauled our driving habits, our dating styles and even our posture. Despite the occasional headlines claiming that digital technology is rotting our brains, not to mention what it’s doing to our children, we’ve welcomed this alluring life partner with open arms and swiping thumbs.

Scientists suspect that these near-constant interactions with digital technology influence our brains. Small studies are turning up hints that our devices may change how we remember, how we navigate and how we create happiness — or not.
Somewhat limited, occasionally contradictory findings illustrate how science has struggled to pin down this slippery, fast-moving phenomenon. Laboratory studies hint that technology, and its constant interruptions, may change our thinking strategies. Like our husbands and wives, our devices have become “memory partners,” allowing us to dump information there and forget about it — an off-loading that comes with benefits and drawbacks. Navigational strategies may be shifting in the GPS era, a change that might be reflected in how the brain maps its place in the world. Constant interactions with technology may even raise anxiety in certain settings.

Yet one large study that asked people about their digital lives suggests that moderate use of digital technology has no ill effects on mental well-being.

The question of how technology helps and hinders our thinking is incredibly hard to answer. Both lab and observational studies have drawbacks. The artificial confines of lab experiments lead to very limited sets of observations, insights that may not apply to real life, says experimental psychologist Andrew Przybylski of the University of Oxford. “This is a lot like drawing conclusions about the effects of baseball on players’ brains after observing three swings in the batting cage.”

Observational studies of behavior in the real world, on the other hand, turn up associations, not causes. It’s hard to pull out real effects from within life’s messiness. The goal, some scientists say, is to design studies that bring the rigors of the lab to the complexities of real life, and then to use the resulting insights to guide our behavior. But that’s a big goal, and one that scientists may never reach.

Evolutionary neurobiologist Leah Krubitzer is comfortable with this scientific ambiguity. She doesn’t put a positive or negative value on today’s digital landscape. Neither good nor bad, it just is what it is: the latest iteration on the continuum of changing environments, says Krubitzer, of the University of California, Davis.

“I can tell you for sure that technology is changing our brains,” she says. It’s just that so far, no one knows what those changes mean.

Of course, nearly everything changes the brain. Musical training reshapes parts of the brain. Learning the convoluted streets of London swells a mapmaking structure in the brains of cabbies. Even getting a good night’s sleep changes the brain. Every aspect of our environment can influence brain and behaviors. In some ways, digital technology is no different. Yet some scientists suspect that there might be something particularly pernicious about digital technology’s grip on the brain.

“We are information-seeking creatures,” says neuroscientist Adam Gazzaley of the University of California, San Francisco. “We are driven to it in very powerful ways.” Today’s digital tools give us unprecedented exposure to information that doesn’t wait for you to seek it out; it seeks you out, he says. That pull is nearly irresistible.

Despite the many unanswered questions about whether our digital devices are influencing our brains and behaviors, and whether for good or evil, technology is galloping ahead. “We should have been asking ourselves [these sorts of questions] in the ’70s or ’80s,” Krubitzer says. “It’s too late now. We’re kind of closing the barn doors after the horses got out.”
Attention grabber
One way in which today’s digital technology is distinct from earlier advances (like landline telephones) is the sheer amount of time people spend with it. In just a decade, smartphones have saturated the market, enabling instant internet access to an estimated 2 billion people around the world. In one small study reported in 2015, 23 adults, ages 18 to 33, spent an average of five hours a day on their phones, broken up into 85 distinct daily sessions. When asked how many times they thought they used their phones, participants underestimated by half.

In a different study, Larry Rosen, a psychologist at California State University, Dominguez Hills, used an app to monitor how often college students unlocked their phones. The students checked their phones an average of 60 times a day, each session lasting about three to four minutes for a total of 220 minutes a day. That’s a lot of interruption, Rosen says.
Smartphones are “literally omnipresent 24-7, and as such, it’s almost like an appendage,” he says. And often, we are compelled to look at this new, alluring rectangular limb instead of what’s around us. “This device is really powerful,” Rosen says. “It’s really influencing our behavior. It’s changed the way we see the world.”

Technology does that. Printing presses, electricity, televisions and telephones all shifted people’s habits drastically, Przybylski says. He proposes that the furor over digital technology melting brains and crippling social lives is just the latest incarnation of the age-old fear of change. “You have to ask yourself, ‘Is there something magical about the power of an LCD screen?’ ” Przybylski says.

Yet some researchers suspect that there is something particularly compelling about this advance. “It just feels different. Computers and the internet and the cloud are embedded in our lives,” says psychologist Benjamin Storm of the University of California, Santa Cruz. “The scope of the amount of information we have at our fingertips is beyond anything we’ve ever experienced. The temptation to become really reliant on it seems to be greater.”

Memory outsourcing
Our digital reliance may encourage even more reliance, at least for memory, Storm’s work suggests. Sixty college undergraduates were given a mix of trivia questions — some easy, some hard. Half of the students had to answer the questions on their own; the other half were told to use the internet. Later, the students were given an easier set of questions, such as “What is the center of a hurricane called?” This time, the students were told they could use the internet if they wanted.

People who had used the internet initially were more likely to rely on internet help for the second, easy set of questions, Storm and colleagues reported online last July in Memory. “People who had gotten used to using the internet continued to do so, even though they knew the answer,” Storm says. This kind of overreliance may signal a change in how people use their memory. “No longer do we just rely on what we know,” he says.
That work builds on results published in a 2011 paper in Science . A series of experiments showed that people who expected to have access to the internet later made less effort to remember things . In this way, the internet has taken the place formerly filled by spouses who remember birthdays, grandparents who remember recipes and coworkers who remember the correct paperwork codes — officially known as “transactive memory partners.”
“We are becoming symbiotic with our computer tools,” Betsy Sparrow, then at Columbia University, and colleagues wrote in 2011. “The experience of losing our internet connection becomes more and more like losing a friend. We must remain plugged in to know what Google knows.”

That digital crutch isn’t necessarily a bad thing, Storm points out. Human memory is notoriously squishy, susceptible to false memories and outright forgetting. The internet, though imperfect, can be a resource of good information. And it’s not clear, he says, whether our memories are truly worse, or whether we perform at the same level, but just reach the answer in a different way.

“Some people think memory is absolutely declining as a result of us using technology,” he says. “Others disagree. Based on the current data, though, I don’t think we can really make strong conclusions one way or the other.”

The potential downsides of this memory outsourcing are nebulous, Storm says. It’s possible that digital reliance influences — and perhaps even weakens — other parts of our thinking. “Does it change the way we learn? Does it change the way we start to put information together, to build our own stories, to generate new ideas?” Storm asks. “There could be consequences that we’re not necessarily aware of yet.”

Research by Gazzaley and others has documented effects of interruptions and multitasking, which are hard to avoid with incessant news alerts, status updates and Instagrams waiting in our pockets. Siphoning attention can cause trouble for a long list of thinking skills, including short- and long-term memory, attention, perception and reaction time. Those findings, however, come from experiments in labs that ask a person to toggle between two tasks while undergoing a brain scan, for instance. Similar effects have not been as obvious for people going about their daily lives, Gazzaley says. But he is convinced that constant interruptions — the dings and buzzes, our own restless need to check our phones — are influencing our ability to think.

Making maps
Consequences of technology are starting to show up for another cognitive task — navigating, particularly while driving. Instead of checking a map and planning a route before a trip, people can now rely on their smartphones to do the work for them. Anecdotal news stories describe people who obeyed the tinny GPS voice that instructed them to drive into a lake or through barricades at the entrance of a partially demolished bridge. Our navigational skills may be at risk as we shift to neurologically easier ways to find our way, says cognitive neuroscientist Véronique Bohbot of McGill University in Montreal.

Historically, getting to the right destination required a person to have the lay of the land, a mental map of the terrain. That strategy takes more work than one that’s called a “response strategy,” the type of navigating that starts with an electronic voice command. “You just know the response — turn right, turn left, go straight. That’s all you know,” Bohbot says. “You’re on autopilot.”
A response strategy is easier, but it leaves people with less knowledge. People who walked through a town in Japan with human guides did a better job later navigating the same route than people who had walked with GPS as a companion, researchers have found.

Scientists are looking for signs that video games, which often expose people to lots of response-heavy situations, influence how people get around. In a small study, Bohbot and colleagues found that people who average 18 hours a week playing action video games such as Call of Duty navigated differently than people who don’t play the games. When tested on a virtual maze, players of action video games were more likely to use the simpler response learning strategy to make their way through, Bohbot and colleagues reported in 2015 in Proceedings of the Royal Society B.

That easier type of response navigation depends on the caudate nucleus, a brain area thought to be involved in habit formation and addiction. In contrast, nerve cells in the brain’s hippocampus help create mental maps of the world and assist in the more complex navigation. Some results suggest that people who use the response method have bigger caudate nuclei, and more brain activity there. Conversely, people who use spatial strategies that require a mental map have larger, busier hippocampi.

Those results on video game players are preliminary and show an association within a group that may share potentially confounding similarities. Yet it’s possible that getting into a habit of mental laxity may change the way people navigate. Digital technology isn’t itself to blame, Bohbot says. “It’s not the technology that’s necessarily good or bad for our brain. It’s how we use the technology,” she says. “We have a tendency to use it in the way that seems to be easiest for us. We’re not making the effort.”

Parts of the brain, including those used to navigate, have many jobs. Changing one aspect of brain function with one type of behavior might have implications for other aspects of life. A small study by Bohbot showed that people who navigate by relying on the addiction-related caudate nucleus smoke more cigarettes, drink more alcohol and are more likely to use marijuana than people who rely on the hippocampus. What to make of that association is still very much up in the air.

Sweating the smartphone
Other researchers are trying to tackle questions of how technology affects our psychological outlooks. Rosen and colleagues have turned up clues that digital devices have become a new source of anxiety for people.
In diabolical experiments, Cal State’s Rosen takes college students’ phones away, under the ruse that the devices are interfering with laboratory measurements of stress, such as heart rate and sweating. The phones are left on, but placed out of reach of the students, who are reading a passage. Then, the researchers start texting the students, who are forced to listen to the dings without being able to see the messages or respond. Measurements of anxiety spike, Rosen has found, and reading comprehension dwindles.

Other experiments have found that heavy technology users last about 10 minutes without their phones before showing signs of anxiety.

Fundamentally, an interruption in smartphone access is no different from those in the days before smartphones, when the landline rang as you were walking into the house with bags full of groceries, so you missed the call. Both situations can raise anxiety over a connection missed. But Rosen suspects that our dependence on digital technology causes these situations to occur much more often.

“The technology is magnificent,” he says. “Having said that, I think that this constant bombardment of needing to check in, needing to be connected, this feeling of ‘I can’t be disconnected, I can’t cut the tether for five minutes,’ that’s going to have a long-term effect.”

The question of whether digital technology is good or bad for people is nearly impossible to answer, but a survey of 120,000 British 15-year-olds (99.5 percent reported using technology daily) takes a stab at it. Oxford’s Przybylski and Netta Weinstein at Cardiff University in Wales have turned up hints that moderate use of digital technology — TV, computers, video games and smartphones — correlates with good mental health, measured by questions that asked about happiness, life satisfaction and social activity.

When the researchers plotted technology use against mental well-being, an umbrella-shaped curve emerged, highlighting what the researchers call the “Goldilocks spot” of technology use — not too little and not too much.

“We found that you’ve got to do a lot of texting before it hurts,” Przybylski says. For smartphone use, the shift from benign to potentially harmful came after about two hours of use on weekdays, mathematical analyses revealed. Weekday recreational computer use had a longer limit: four hours and 17 minutes, the researchers wrote in the February Psychological Science.
For even the heaviest users, the relationship between technology use and poorer mental health wasn’t all that strong. For scale, the potential negative effects of all that screen time was less than a third of the size of the positive effects of eating breakfast, Przybylski and Weinstein found.

Even if a relationship is found between technology use and poorer mental health, scientists still wouldn’t know why, Przybylski says. Perhaps the effect comes from displacing something, such as exercise or socializing, and not the technology itself.

We may never know just how our digital toys shape our brains. Technology is constantly changing, and fast. Our brains are responding and adapting to it.

“The human neocortex basically re-creates itself over successive generations,” Krubitzer says. It’s a given that people raised in a digital environment are going to have brains that reflect that environment. “We went from using stones to crack nuts to texting on a daily basis,” she says. “Clearly the brain has changed.”

It’s possible that those changes are a good thing, perhaps better preparing children to succeed in a fast-paced digital world. Or maybe we will come to discover that when we no longer make the effort to memorize our best friend’s phone number, something important is quietly slipping away.

It’s time to redefine what qualifies as a planet, scientists propose

Pluto is a planet. It always has been, and it always will be, says Will Grundy of Lowell Observatory in Flagstaff, Arizona. Now he just has to convince the world of that.

For centuries, the word planet meant “wanderer” and included the sun, the moon, Mercury, Venus, Mars, Jupiter and Saturn. Eventually the moon and sun were dropped from the definition, but Pluto was included, after its discovery in 1930. That idea of a planet as a rocky or gaseous body that orbited the sun stuck, all the way up until 2006.
Then, the International Astronomical Union narrowed the definition, describing a planet as any round object that orbits the sun and has moved any pesky neighbors out of its way, either by consuming them or flinging them off into space. Pluto failed to meet the last criterion (SN: 9/2/06, p. 149), so it was demoted to a dwarf planet.

Almost overnight, the solar system was down to eight planets. “The public took notice,” Grundy says. It latched onto the IAU’s definition — perhaps a bit prematurely. The definition has flaws, he and other planetary scientists argue. First, it discounts the thousands of exotic worlds that orbit other stars and also rogue ones with no star to call home (SN: 4/4/15, p. 22).

Second, it requires that a planet cut a clear path around the sun. But no planet does that; Earth, Mars, Jupiter and Neptune share their paths with asteroids, and objects crisscross planets’ paths all the time.

The third flaw is related to the second. Objects farther from the sun need to be pretty bulky to cut a clear path. You could have a rock the size of Earth in the Kuiper Belt and it wouldn’t have the heft required to gobble down or eject objects from its path. So, it couldn’t be considered a planet.

Grundy and colleagues (all members of NASA’s New Horizons mission to Pluto) laid out these arguments against the IAU definition of a planet March 21 at the Lunar and Planetary Science Conference in The Woodlands, Texas.
A more suitable definition of a planet, says Grundy, is simpler: It’s any round object in space that is smaller than a star. By that definition, Pluto is a planet. So is the asteroid-belt object Ceres. So is Earth’s moon. “There’d be about 110 known planets in our solar system,” Grundy says, and plenty of exoplanets and rogue worlds would fit the bill as well.

The reason for the tweak is to keep the focus on the features — the physics, the geology, the atmosphere — of the world itself, rather than worry about what’s going on around it, he says.

The New Horizons mission has shown that Pluto is an interesting world with active geology, an intricate atmosphere and other features associated with planets in the solar system. It makes no sense to write Pluto off because it doesn’t fit one criterion. Grundy seems convinced the public could easily readopt the small world as a planet. Though he admits astronomers might be a tougher sell.

“People have been using the word correctly all along,” Grundy says. He suggests we stick with the original definition. That’s his plan.

CRISPR had a life before it became a gene-editing tool

It is the dazzling star of the biotech world: a powerful new tool that can deftly and precisely alter the structure of DNA. It promises cures for diseases, sturdier crops, malaria-resistant mosquitoes and more. Frenzy over the technique — known as CRISPR/Cas9 — is in full swing. Every week, new CRISPR findings are unfurled in scientific journals. In the courts, universities fight over patents. The media report on the breakthroughs as well as the ethics of this game changer almost daily.

But there is a less sequins-and-glitter side to CRISPR that’s just as alluring to anyone thirsty to understand the natural world. The biology behind CRISPR technology comes from a battle that has been raging for eons, out of sight and yet all around us (and on us, and in us).

The CRISPR editing tool has its origins in microbes — bacteria and archaea that live in obscene numbers everywhere from undersea vents to the snot in the human nose. For billions of years, these single-celled organisms have been at odds with the viruses — known as phages — that attack them, invaders so plentiful that a single drop of seawater can hold 10 million. And natural CRISPR systems (there are many) play a big part in this tussle. They act as gatekeepers, essentially cataloging viruses that get into cells. If a virus shows up again, the cell — and its offspring — can recognize and destroy it. Studying this system will teach biologists much about ecology, disease and the overall workings of life on Earth.

But moving from the simple, textbook story into real life is messy. In the few years since the defensive function of CRISPR systems was first appreciated, microbiologists have busied themselves collecting samples, conducting experiments and crunching reams of DNA data to try to understand what the systems do. From that has come much elegant physiology, a mass of complexity, surprises aplenty — and more than a little mystery.
Spoiled yogurt
The biology is complicated, and its basic nuts and bolts took some figuring out. There are two parts to CRISPR/Cas systems: the CRISPR bit and the Cas bit. The CRISPR bit — or “clustered regularly interspaced short palindromic repeats” — was stumbled on in the late 1980s and 1990s. Scientists then slowly pieced the story together by studying microbes that thrive in animals’ guts and in salt marshes, that cause the plague and that are used to make delicious yogurt and cheese.

None of the scientists knew what they were dealing with at first. They saw stretches of DNA with a characteristic pattern: short lengths of repeated sequence separated by other DNA sequences now known as spacers. Each spacer was unique. Because the roster of spacers could differ from one cell to the next in a given microbe species, an early realization was that these differences could be useful for forensic “typing” — investigators could tell whether food poisoning cases were linked, or if someone had stolen a company’s yogurt starter culture.
But curious findings piled up. Some of those spacers, it turned out, matched the DNA of phages. In a flurry of reports in 2005, scientists showed, to name one example, that strains of the lactic acid bacterium Streptococcus thermophilus contained spacers that matched genetic material of phages known to infect Streptococcus. And the more spacers a strain had, the more resistant it was to attack by phages.

This began to look a lot like learned or adaptive immunity, akin to our own antibody system: After exposure to a specific threat, your immune system remembers and you are thereafter resistant to that threat. In a classic experiment published in Science in 2007, researchers at the food company Danisco showed it was so. They could see new spacers added when a phage infected a culture of S. thermophilus. Afterward, the bacterium was immune to the phage. They could artificially engineer a phage spacer into the CRISPR DNA and see resistance emerge; when they took the spacer away, immunity was lost.

This was handy intel for an industry that could find whole vats of yogurt-making bacteria wiped out by phage infestations. It was an exciting time scientifically and commercially, says Rodolphe Barrangou of North Carolina State University in Raleigh, who did a lot of the Danisco work. “It was not just discovering a cool system, but also uncovering a powerful phage-resistance technology for the dairy industry,” he says.

The second part of the CRISPR/Cas system is the Cas bit: a set of genes located near the cluster of CRISPR spacers. The DNA sequences of these genes strongly suggested that they carried instructions for proteins that interact with DNA or RNA in some fashion — sticking to it, cutting it, copying it, unraveling it. When researchers inactivated one Cas gene or another, they saw immunity falter. Clearly, the two bits of the system — CRISPR and Cas — were a team.
It took many more experiments to get to today’s basic model of how CRISPR/Cas systems fight phages — and not just phages. Other types of foreign DNA can get into microbes, including circular rings called plasmids that shuttle from cell to cell and DNA pieces called transposable elements, which jump around within genomes. CRISPRs can fend off these intruders, as well as keep a microbe’s genome in tidy order.

The process works like this: A virus injects its genetic material into the cell. Sensing this danger, the cell selects a little strip of that genetic material and adds it to the spacers in the CRISPR cluster. This step, known as immunization or adaptation, creates a list of encounters a cell has had with viruses, plasmids or other foreign bits of DNA over time — neatly lined up in reverse chronological order, newest to oldest.

Older spacers eventually get shed, but a CRISPR cluster can grow to be long — the record holder to date is 587 spacers in Haliangium ochraceum, a salt-loving microbe isolated from a piece of seaweed. “It’s like looking at the last 600 shots you had in your arm,” says Barrangou. “Think about that.”

New spacer in place, the microbe is now immunized. Later comes targeting. If that same phage enters the cell again, it’s recognized. The cell has made RNA copies of the relevant spacer, which bind to the matching spot on the genome of the invading phage. That “guide RNA” leads Cas proteins to target and snip the phage DNA, defanging the intruder.
Researchers now know there are a confetti-storm of different CRISPR systems, and the list continues to grow. Some are simple — such as the CRISPR/Cas9 system that’s been adapted for gene editing in more complex creatures (SN: 4/15/17, p. 16) — and some are elaborate, with many protein workhorses deployed to get the job done.

Those who are sleuthing the evolution of CRISPR systems are deciphering a complex story. The part of the CRISPR toolbox involved in immunity (adding spacers after phages inject their genetic material) seems to have originated from a specific type of transposable element called a casposon. But the part responsible for targeting has multiple origins — in some cases, it’s another type of transposable element. In others, it’s a mystery.

The downsides
Given the power of CRISPR systems to ward off foes, one might think every respectable microbe out there in the soils, vents, lakes, guts and nostrils of this planet would have one. Not so.

Numbers are far from certain, partly because science hasn’t come close to identifying all the world’s microbes, let alone probe them all for CRISPRs. But the scads of microbial genetic data accrued so far throw up interesting trends.

Tallies suggest that CRISPR systems are far more prevalent in known archaea than in known bacteria — such systems exist in roughly 90 percent of archaea and about 35 percent of bacteria, says Eugene Koonin, a computational evolutionary biologist at the National Institutes of Health in Bethesda, Md. Archaea and bacteria, though both small and single-celled, are on opposite sides of the tree of life.

Perhaps more significantly, Koonin says, almost all the known microbes that live in superhot environments have CRISPRs. His group’s math models suggest that CRISPR systems are most useful when microbes encounter a big enough variety of viruses to make adaptive memory worth having. But if there’s too much variety, and viruses are changing very fast, CRISPRs don’t really help — because you’d never see the same virus again. The superhot ecosystems, he says, seem to have a stable amount of phage diversity that’s not too high or low.

And CRISPR systems have downsides. Just as people can develop autoimmune reactions against their own bodies, bacteria and archaea can accidentally make CRISPR spacers from bits of their own DNA — and risk chewing up their own genetic material. Researchers have seen this happen. “No immunity comes without a cost,” says Rotem Sorek, a microbial genomicist at the Weizmann Institute of Science in Rehovot, Israel.

But mistakes are rare, and Sorek and his colleagues recently figured out why in the microbe they study. The researchers reported in Nature in 2015 that CRISPR spacers are created from linear bits of DNA — and phage DNA is linear when it enters cells. The bacterial chromosome is protected because of its circular form. Should it break and become linear for a spell, such as when it’s being replicated, it contains signals that ward off the Cas proteins.

There are other negatives to CRISPR systems. It’s not always a bonus to keep out phages and other invaders, which can sometimes bring in useful things. Escherichia coli O157:H7, of food poisoning fame, can make humans sick because of toxin genes it harbors that were brought in by a phage, to name just one of myriad examples. Even CRISPR systems themselves are spread around the microbial kingdom via phages, plasmids or transposable elements.

For microbes that lack CRISPR systems, there are many other ways to repel foreign DNA — as much as 10 percent of a microbial genome may be devoted to hawkish warfare, and new defense systems are still being uncovered.

Countermeasures
The war between bacteria and phages is two-sided, of course. Just as a microbe wants to keep doors shut to protect its genetic integrity and escape destruction, the phage wants in.

And so the phage fights back against CRISPRs. It genetically morphs into forms that CRISPRs no longer recognize. Or it designs bespoke artillery. Microbiologist Joe Bondy-Denomy, now at the University of California, San Francisco, happened upon such customized weapons as a grad student in the lab of molecular microbiologist Alan Davidson at the University of Toronto. The team knew that the bacterium Pseudomonas aeruginosa, which lives in soil and water and can cause dangerous infections, has a vigorous CRISPR system. Yet some phages didn’t seem fazed by it.

That’s because those phages have small proteins that will bind to and interfere with this or that part of the CRISPR machinery, such as the Cas enzyme that cuts phage DNA. The binding disables the CRISPR system, the researchers reported in 2015 in Nature. Bondy-Denomy and others have since found anti-CRISPR genes in other phages and other kinds of interloping DNA. The genes are so common, Davidson says, that he wonders how many CRISPR systems are truly active.

In an especially bizarre twist, microbiologist Kimberley Seed of the University of California, Berkeley found a phage that carries its own CRISPR system and uses it to fight back against the cholera bacterium it invades, she and colleagues reported in 2013 in Nature. It chops up a segment of bacterial DNA that normally inhibits phage infection.

Of course, in this never-ending scuffle one would expect the microbes to again fight back against the phages. “It’s something I often get asked: ‘Great, the anti-CRISPRs are there, so where are the anti-anti-CRISPRs?’ ” Bondy-Denomy says. Nobody has found such things yet.

Evolution drivers
It’s one thing to study CRISPR systems in well-controlled lab settings, or in just one type of microbe. It’s another to understand what all the various CRISPRs do to shape the ecosystem of a bubbling hot spring, human gut, diseased lung or cholera-tainted river. Estimates of CRISPR abundance could drop as more sampling is done, especially of dark horse microbes that researchers know little about.

In a 2016 report in Nature Communications, for example, geomicrobiologist Jill Banfield of UC Berkeley and colleagues detected 1,724 microbes in Colorado groundwater that had been treated to boost the abundance of types that are difficult to isolate. CRISPR systems were much rarer in this sample than in databases of better-known microbes.

Tallying CRISPRs is just the start, of course. Microbial communities — including those inside our own guts, where there are plenty of CRISPR systems and phages — are dynamic, not frozen. How do CRISPRs shape the evolution of phages and microbes in the wild? Banfield’s and Barrangou’s labs teamed up to watch as S. thermophilus and phages incubated together in a milk medium for hundreds of days. The team saw bacterial numbers fall as phages invaded; then bacteria acquired spacers against the phage and rallied — and phage numbers fell downward in turn. Then new phage populations sprang up, immune to S. thermophilus defenses because of genetic changes. In this way, the researchers reported in 2016 in mBio, CRISPRs are “one of the fundamental drivers of phage evolution.”

CRISPR systems can be picked up, dropped, then picked up again by bacteria and archaea over time, perhaps as conditions and needs change. The bacterium Vibrio cholerae is an example of this dynamism, as Seed and colleagues reported in 2015 in the Journal of Bacteriology. The older, classical strains of this medical blight harbored CRISPRs, but these strains went largely extinct in the wild in the 1960s. Strains that cause cholera today do not have CRISPRs.

Nobody knows why, Seed says. But scientists stress that it is a mischaracterization to paint the relationship between microbes and phages, plasmids and transposable elements as a simplistic war. Phages don’t always wreak havoc; they can slip their genomes quietly into the bacterial chromosome and coexist benignly, getting copied along with the host DNA. Phages, plasmids and transposable elements can confer new, useful traits — sometimes even essential ones. Indeed, such movement of DNA across species and strains is at the heart of how bacteria and archaea evolve.

So it’s about finding balance. “If you incorporate too much foreign DNA, you cannot maintain a species,” says Luciano Marraffini, a molecular microbiologist at the Rockefeller University in New York City whose work first showed that DNA-cutting was key to CRISPR systems. But you do need to let some DNA in, and it’s likely that some CRISPR systems permit this: The system he studies in Staphylococcus epidermidis, for example, only goes after phages that are in their cell-killing, or lytic, state, he and colleagues reported in 2014 in Nature.
Beyond defense
One thing is very clear about CRISPR systems: They are perplexing in many ways. For a start, the spacers in a microbe should reflect its own, individual story of the phages it has encountered. So you’d think there would be local pedigrees, that a bacterium sampled in France would have a different spacer cluster from a bacterium sampled in Argentina. This is not what researchers always see.

Take the nasty P. aeruginosa. Rachel Whitaker, a microbial population biologist at the University of Illinois at Urbana-Champaign, studies Pseudomonas samples collected from people with cystic fibrosis, whose lungs develop chronic infections. She’s found no sign that two patients living close to each other carry more-similar P. aeruginosa CRISPRs than two patients thousands of miles apart. Yet surely one would expect nearby CRISPRs to be closer matches, because the Pseudomonas would have encountered similar phages. “It’s very weird,” Whitaker says.

Others have seen the same thing in heat-loving bacteria sampled from very distant bubbling hot springs. It’s as if scientists don’t truly understand how bacteria spread around the world — there could be a strong effect of far-flung passage by air or wind, says Konstantin Severinov, who studies CRISPR systems at Rutgers University in New Brunswick, N.J.
Another weirdness is the differing vigor of CRISPR systems. Some are very active. Molecular biologist Devaki Bhaya of the Carnegie Institution for Science’s plant biology department at Stanford University sees clear signs that spacers are frequently added and dropped in the cyanobacteria of Yellowstone’s hot springs, for example. But other systems are sluggish, and E. coli, that classic workhorse of genetics research, has a respectable-looking CRISPR system — that is switched off.

It may have been off for a long time. Some 42,000 years ago, a baby woolly mammoth died in what is now northwestern Siberia. The remains, found in 2007, were so well-preserved that the intestines were intact and E. coli DNA could be extracted.

In research published in Molecular Ecology in January, Severinov’s team found surprising similarities between the spacers in the mammoth-derived E. coli CRISPR cluster and those in modern-day E. coli. “There was no turnover in all that time,” Severinov marvels. If the CRISPR system isn’t active, why does E. coli bother to keep it?

That quandary leads neatly to what some researchers refer to as an intellectually “scandalous situation.”

In some cases, the genetic sequence of spacers nicely matches phage DNA. But overall, only a fraction (around 1 to 2 percent) of the spacers scientists know about have been matched to a virus or a plasmid. In E. coli, the spacers don’t match common, classic phages known to infect the bacterium. “Is it the case that there is a huge, unknown amount of viral dark matter in the world?” says Koonin — or are phages evolving superfast? “Or is it something completely different?”

Faced with this conundrum, some researchers strongly suspect — and have evidence — that CRISPR systems may do more than defend; they may have other jobs. Communication, perhaps. Or turning genes on and off.

But some microbes’ CRISPR sequences do make sense, especially if looking at the spacers most recently added, and others may be clues to phages still undiscovered. So even as they scratch their heads about many things CRISPR, scientists are also excited by the stories CRISPR clusters can tell about the viruses and other bits of DNA that bacteria and archaea encounter and that they choose, for whatever reason, to note for the record. What do microbes pay attention to? What do they ignore?

CRISPRs offer a bright new window on such questions and, indeed, already are unearthing novel phages and facts about who infects whom in the microscopic world.

“We can catalog everything that’s out there. But we don’t really know what matters,” says Bondy-Denomy. “CRISPRs can help us understand.”