Skip to main content

Scientists tracked billions of migrating ocean animals using a giant space laser

NASA

It sounds like the evil plan of a James Bond villain. An orbiting satellite, with the ever-so-slightly sinister name CALIPSO, is blasting lasers into the Earth’s oceans. If movies have taught us anything, it’s that what follows is a ransom message to the world’s leaders, promising some kind of fatal accident unless they pay up an extortionate ransom to stop it. Well, not quite. In reality, the Cloud-Aerosol Lidar and Infrared Pathfinder Satellite Observations (CALIPSO) satellite has indeed been orbiting the Earth since 2006, and is indeed sending down constant bursts of laser beams from space.

But it’s not to boil the world’s oceans or cause some kind of devastating tidal wave that only Daniel Craig can stop. Instead, the satellite — the work of NASA and French space agency, the Center National d’Etudes Spatiales (CNES) — is being used by scientists to carry out an unprecedented, decade-long study into the world’s largest animal migration. Which, as Mike Behrenfeld, the project lead and senior research scientist and professor at Oregon State University, points out, just so happens to take place in the ocean.

“The migration is made up of countless numbers of zooplankton, which are marine animals consisting of squid, small fish, and things like that,” Behrenfeld told Digital Trends. “They hide in the dark, deep ocean during the day to avoid visual predators. Then every night they swim up to the surface to feed on phytoplankton (read: microscopic marine algae which provide food for sea creatures including whales, shrimp, snails, and jellyfish). When the sun comes up, they swim down to the deepest ocean again to hide.”

The power of space lidar

This migration pattern, called diurnal vertical migration, has been known to science for around 200 years. Ships which used nets to trawl night and day observed vastly different quantities of plankton depending on the time. During World War II, the United States Navy noticed strange sonar readings which they initially thought might be enemy submarines. They later discovered, with help from biologists, that it was the result of the dense movement of large numbers of marine animals. However, it’s never been possible to observe these changes over a long duration before.

That’s where the use of space lidar comes in. A lidar system employs a sensor and a series of laser pulses to measure reflected light. Its most widely publicized application is in self-driving cars, where it helps vehicles build up a three-dimensional representation of the world around them. However, atmospheric lidar has been used for decades to measure atmospheric details such as aerosol particles, ice crystals, water vapor, and more. This is what the CALIPSO satellite was originally developed to do: to measure clouds and aerosol.

NASA

Behrenfeld and his colleagues had a different idea, though. “What we [used it for was to look] at the difference in the optical properties measured by the instrument between day and night,” he said. “The idea is that the light-scattering signal that we measured during the day is going to reflect the organisms that are there all the time. The difference between the day and night signal is the addition of these migrating animals.”

Behrenfeld notes that plenty of lidar systems have done ground-based observations using systems rigged to airplanes or helicopters. In this capacity, lidar has been utilized for everything from studying melting polar ice caps to finding lost cities beneath dense jungle cover.

“But looking at the ocean using space lidar?” he said. “That’s pretty brand new. The beauty of the satellite is that you get global coverage, pole to pole, every 16 days. You simply can’t come close to that with an airplane, a ship, or any other method. It was the power of this global coverage that allowed us to do something very different to what had been done before: to look at the pattern of these things globally and how they change everywhere on the planet, from month to month, for ten years. There’s no other way to do that.”

“To look at the pattern of these things globally and how they change everywhere on the planet, from month to month, for ten years. There’s no other way to do that.”

The team’s work (published in Nature) uncovered some fascinating insights, such as that there are fewer vertically migrating animals in lower-nutrient and clearer waters. However, the use of lidar to measure this biological conveyor belt of animals from the deep sea may be of interest to more than just marine biologists. It also creates a pathway for moving photosynthetically-captured carbon dioxide from the surface to the longer-term carbon storage pools at depth.

During the day, phytoplankton photosynthesize and absorb massive amounts of carbon dioxide. This contributes to the ocean’s ability to absorb atmospheric greenhouse gases. When it is eaten by animals which then defecate it at depths where it is trapped in the ocean, it prevents it from being released back into the atmosphere. Understanding this could therefore be of interest to climate modelers.

Taking things to the next level

Going forward, Behrenfeld hopes this will be just a proof-of-concept for additional more detailed work. “The instrument that we used for this study was not designed for oceans at all. We just piggybacked off something that was already up there to showcase the idea,” he said. “There are definitely better technologies that could give us a ton more information about ocean ecosystems if we could build one of these things for actually doing ocean study.”

For instance, the current technology used is able to probe only the top 22 meters of the ocean (a depth of around 70 feet). But while that’s fine for capturing clouds at the right resolution, it’s not sufficient for the kind of high quality data that Behrenfeld would ideally like to achieve.

“You can make lidar measurements that have centimeter vertical resolution,” he said. “For the ocean, what would be really great would be something that can give us information on the vertical structure of plankton every few meters, down to some relatively deep depths like the depth that sunlight penetrates. That would be amazing.”

The technology does exist. NASA’s Langley Research Center has demonstrated how something called High Spectral Resolution Lidar (HSRL) can be used for oceanic research — although this has yet to be carried out using a space lidar. When this next generation of oceanic laser scanning does take place, Behrenfeld is excited about the possibilities. After all, who knows what we’ll find down there?

Editors' Recommendations

Luke Dormehl
I'm a UK-based tech writer covering Cool Tech at Digital Trends. I've also written for Fast Company, Wired, the Guardian…
Digital Trends’ Top Tech of CES 2023 Awards
Best of CES 2023 Awards Our Top Tech from the Show Feature

Let there be no doubt: CES isn’t just alive in 2023; it’s thriving. Take one glance at the taxi gridlock outside the Las Vegas Convention Center and it’s evident that two quiet COVID years didn’t kill the world’s desire for an overcrowded in-person tech extravaganza -- they just built up a ravenous demand.

From VR to AI, eVTOLs and QD-OLED, the acronyms were flying and fresh technologies populated every corner of the show floor, and even the parking lot. So naturally, we poked, prodded, and tried on everything we could. They weren’t all revolutionary. But they didn’t have to be. We’ve watched enough waves of “game-changing” technologies that never quite arrive to know that sometimes it’s the little tweaks that really count.

Read more
Digital Trends’ Tech For Change CES 2023 Awards
Digital Trends CES 2023 Tech For Change Award Winners Feature

CES is more than just a neon-drenched show-and-tell session for the world’s biggest tech manufacturers. More and more, it’s also a place where companies showcase innovations that could truly make the world a better place — and at CES 2023, this type of tech was on full display. We saw everything from accessibility-minded PS5 controllers to pedal-powered smart desks. But of all the amazing innovations on display this year, these three impressed us the most:

Samsung's Relumino Mode
Across the globe, roughly 300 million people suffer from moderate to severe vision loss, and generally speaking, most TVs don’t take that into account. So in an effort to make television more accessible and enjoyable for those millions of people suffering from impaired vision, Samsung is adding a new picture mode to many of its new TVs.
[CES 2023] Relumino Mode: Innovation for every need | Samsung
Relumino Mode, as it’s called, works by adding a bunch of different visual filters to the picture simultaneously. Outlines of people and objects on screen are highlighted, the contrast and brightness of the overall picture are cranked up, and extra sharpness is applied to everything. The resulting video would likely look strange to people with normal vision, but for folks with low vision, it should look clearer and closer to "normal" than it otherwise would.
Excitingly, since Relumino Mode is ultimately just a clever software trick, this technology could theoretically be pushed out via a software update and installed on millions of existing Samsung TVs -- not just new and recently purchased ones.

Read more
AI turned Breaking Bad into an anime — and it’s terrifying
Split image of Breaking Bad anime characters.

These days, it seems like there's nothing AI programs can't do. Thanks to advancements in artificial intelligence, deepfakes have done digital "face-offs" with Hollywood celebrities in films and TV shows, VFX artists can de-age actors almost instantly, and ChatGPT has learned how to write big-budget screenplays in the blink of an eye. Pretty soon, AI will probably decide who wins at the Oscars.

Within the past year, AI has also been used to generate beautiful works of art in seconds, creating a viral new trend and causing a boon for fan artists everywhere. TikTok user @cyborgism recently broke the internet by posting a clip featuring many AI-generated pictures of Breaking Bad. The theme here is that the characters are depicted as anime characters straight out of the 1980s, and the result is concerning to say the least. Depending on your viewpoint, Breaking Bad AI (my unofficial name for it) shows how technology can either threaten the integrity of original works of art or nurture artistic expression.
What if AI created Breaking Bad as a 1980s anime?
Playing over Metro Boomin's rap remix of the famous "I am the one who knocks" monologue, the video features images of the cast that range from shockingly realistic to full-on exaggerated. The clip currently has over 65,000 likes on TikTok alone, and many other users have shared their thoughts on the art. One user wrote, "Regardless of the repercussions on the entertainment industry, I can't wait for AI to be advanced enough to animate the whole show like this."

Read more