Close Menu
    Facebook X (Twitter) Instagram
    SciTechDaily
    • Biology
    • Chemistry
    • Earth
    • Health
    • Physics
    • Science
    • Space
    • Technology
    Facebook X (Twitter) Pinterest YouTube RSS
    SciTechDaily
    Home»Technology»How Scientists Manage the Flood of “Big Data” from Space
    Technology

    How Scientists Manage the Flood of “Big Data” from Space

    By Whitney Clavin, Jet Propulsion LaboratoryOctober 21, 20132 Comments5 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn WhatsApp Email Reddit
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email Reddit
    How Scientists Manage the Deluge of Big Data from Space
    The center of the Milky Way galaxy imaged by NASA’s Spitzer Space Telescope is displayed on a quarter-of-a-billion-pixel, high-definition 23-foot-wide (7-meter) LCD science visualization screen at NASA’s Ames Research Center in Moffett Field, California. Credit: NASA/Ames/JPL-Caltech

    As “big data” from space missions continues to pour in, scientists and software engineers are coming up with new strategies for managing the ever-increasing flow of such large and complex data streams.

    For NASA and its dozens of missions, data pour in every day like rushing rivers. Spacecraft monitor everything from our home planet to faraway galaxies, beaming back images and information to Earth. All those digital records need to be stored, indexed, and processed so that spacecraft engineers, scientists, and people across the globe can use the data to understand Earth and the universe beyond.

    At NASA’s Jet Propulsion Laboratory in Pasadena, California, mission planners and software engineers are coming up with new strategies for managing the ever-increasing flow of such large and complex data streams, referred to in the information technology community as “big data.”

    How big is big data? For NASA missions, hundreds of terabytes are gathered every hour. Just one terabyte is equivalent to the information printed on 50,000 trees worth of paper.

    “Scientists use big data for everything from predicting weather on Earth to monitoring ice caps on Mars to searching for distant galaxies,” said Eric De Jong of JPL, principal investigator for NASA’s Solar System Visualization project, which converts NASA mission science into visualization products that researchers can use. “We are the keepers of the data, and the users are the astronomers and scientists who need images, mosaics, maps, and movies to find patterns and verify theories.”

    Building Castles of Data

    De Jong explains that there are three aspects to wrangling data from space missions: storage, processing and access. The first task, to store or archive the data, is naturally more challenging for larger volumes of data. The Square Kilometer Array (SKA), a planned array of thousands of telescopes in South Africa and Australia, illustrates this problem. Led by the SKA Organization based in England and scheduled to begin construction in 2016, the array will scan the skies for radio waves coming from the earliest galaxies known.

    JPL is involved with archiving the array’s torrents of images: 700 terabytes of data are expected to rush in every day. That’s equivalent to all the data flowing on the Internet every two days. Rather than build more hardware, engineers are busy developing creative software tools to better store the information, such as “cloud computing” techniques and automated programs for extracting data.

    “We don’t need to reinvent the wheel,” said Chris Mattmann, a principal investigator for JPL’s big-data initiative. “We can modify open-source computer codes to create faster, cheaper solutions.” Software that is shared and free for all to build upon is called open source or open code. JPL has been increasingly bringing open-source software into its fold, creating improved data processing tools for space missions. The JPL tools then go back out into the world for others to use for different applications.

    “It’s a win-win solution for everybody,” said Mattmann.

    In Living Color

    Archiving isn’t the only challenge in working with big data. De Jong and his team develop new ways to visualize the information. Each image from one of the cameras on NASA’s Mars Reconnaissance Orbiter, for example, contains 120 megapixels. His team creates movies from data sets like these, in addition to computer graphics and animations that enable scientists and the public to get up close with the Red Planet.

    “Data are not just getting bigger but more complex,” said De Jong. “We are constantly working on ways to automate the process of creating visualization products, so that scientists and engineers can easily use the data.”

    Data Served Up to Go

    Another big job in the field of big data is making it easy for users to grab what they need from the data archives.

    “If you have a giant bookcase of books, you still have to know how to find the book you’re looking for,” said Steve Groom, manager of NASA’s Infrared Processing and Analysis Center at the California Institute of Technology, Pasadena. The center archives data for public use from a number of NASA astronomy missions, including the Spitzer Space Telescope, the Wide-field Infrared Survey Explorer (WISE), and the U.S. portion of the European Space Agency’s Planck mission.

    Sometimes users want to access all the data at once to look for global patterns, a benefit of big data archives. “Astronomers can also browse all the ‘books’ in our library simultaneously, something that can’t be done on their own computers,” said Groom.

    “No human can sort through that much data,” said Andrea Donnellan of JPL, who is charged with a similarly mountainous task for the NASA-funded QuakeSim project, which brings together massive data sets – space- and Earth-based – to study earthquake processes.

    QuakeSim’s images and plots allow researchers to understand how earthquakes occur and develop long-term preventative strategies. The data sets include GPS data for hundreds of locations in California, where thousands of measurements are taken, resulting in millions of data points. Donnellan and her team develop software tools to help users sift through the flood of data.

    Ultimately, the tide of big data will continue to swell, and NASA will develop new strategies to manage the flow. As new tools evolve, so will our ability to make sense of our universe and the world.

    Never miss a breakthrough: Join the SciTechDaily newsletter.
    Follow us on Google and Google News.

    Astronomy JPL NASA
    Share. Facebook Twitter Pinterest LinkedIn Email Reddit

    Related Articles

    The Incredible Story of “The Camera That Saved Hubble”

    ‘Chemical Laptop’ Could Be Used to Search for Life Beyond Earth

    Superconducting Amplifier Designed to Study the Universe

    Sombrero Galaxy Exhibits Dual Characteristics

    Searching for “Bubbles” in the Milky Way

    NASA’s Spitzer Space Telescope Finds Solid Buckyballs in Space

    SCaN Testbed Headed for International Space Station Later This Year

    NASA’s GRAIL Mission Returns Video From Far Side of the Moon

    NASA’s NuSTAR is One Step Closer to Launching

    2 Comments

    1. Arturo Rojas on October 26, 2013 9:44 am

      I would like how NASA drive the amount of data from space, space missions and how they hope to work with Big Data.

      Reply
    2. N. Coppedge on December 7, 2013 7:13 pm

      Sometime in the near future it might be relevant to map coherent qualities using an axiometric as opposed to axonometric approach. Perhaps this new model is in fact a secondary approach which is better situated in designing theoretical environments than in its applications to materials which have already been abbreviated to fit flexible models. However, ultimately axiometry such as ‘A, B, C, D = AB-CD or AD-CB’ when all terms are opposites may be relevant for abbreviating the specific meaning of data, and in collecting any type of information into aggregates for deeper systems.

      Methods for categorical deduction, which is the central method, I believe were only introduced recently, in my book, The Dimensional Philosopher’s Toolkit. Although it may not be calculus, I believe that its implications are wide-ranging, and ultimately could serve to make complex models more effective, by associating qualities with quantities, or providing a theoretical background for any new assumptions, including unknown data and the relevance of existing models.

      There is a role for geniuses is probably a bygone conclusion of astronomy these days, but if there is any question of missing piths and indexes which could add to the value of computing, I suggest looking in my encyclopedia.

      Reply
    Leave A Reply Cancel Reply

    • Facebook
    • Twitter
    • Pinterest
    • YouTube

    Don't Miss a Discovery

    Subscribe for the Latest in Science & Tech!

    Trending News

    Scientists Break 50-Year-Old Bottleneck To Supercharge Cancer Drug Production

    Popular Weight-Loss Drug Wegovy Linked to Sudden Vision Loss

    Extraordinary Fossil Reveals Fatal Duel Between Ocean Titans 80 Million Years Ago

    “Super Bizarre” – Neuroscientists Discover That Adult Brain Is Filled With Millions of “Silent Synapses”

    Simple Brain Training Cuts Dementia Risk Decades Later, Study Finds

    A Simple Injection Could Help the Heart Heal Itself After a Heart Attack

    Scientists Just Discovered a Hidden Freshwater World Beneath the Great Salt Lake

    Why Your Daily Shower Could Be Worsening the Water Crisis

    Follow SciTechDaily
    • Facebook
    • Twitter
    • YouTube
    • Pinterest
    • Newsletter
    • RSS
    SciTech News
    • Biology News
    • Chemistry News
    • Earth News
    • Health News
    • Physics News
    • Science News
    • Space News
    • Technology News
    Recent Posts
    • Scientists Debunk Rattlesnake Myth That Fooled Hikers and Doctors for Decades
    • This Bizarre Insect Turns Pink to Green in Just 11 Days
    • Scientists Discover Plants Can “Count” – and May Be Smarter Than We Thought
    • New Study Challenges the Idea That We Stop Psychologically Growing After 30
    • Primordial Magnetic Fields May Solve One of Cosmology’s Biggest Mysteries
    Copyright © 1998 - 2026 SciTechDaily. All Rights Reserved.
    • Science News
    • About
    • Contact
    • Editorial Board
    • Privacy Policy
    • Terms of Use

    Type above and press Enter to search. Press Esc to cancel.