Sometime in the 2010s, if all goes well, the Large Synoptic Survey Telescope (LSST) will start to bring a vision of the heavens to Earth. Suspended between its vast mirrors will be a three-billion-pixel sensor array, which on a clear winter night will produce 30 terabytes of data. In less than a week this remarkable telescope will map the whole night sky with a greater speed and sensitivity than could have been imagined more than a decade or so ago, recapitulating with added detail the entire history of optical astronomy from Galileo to the Palomar Sky Survey.

And then the next week it will do the same again, looking for transient changes, adding new information and building up a database of billions of objects and millions of billions of bytes.

When looking at the future of scientific computing, as Nature does this week in a selection of News Features and Commentaries (starting on pages 398 and 409), it is easy to focus on the vast data architectures necessary for projects such as the LSST or the Large Hadron Collider at CERN, the European particle-physics laboratory near Geneva. The truly amazing story, though, is of the distributed power that ends up not in exceptional places such as the focal plane of a giant telescope, but spread out across the world; the power that allows data to be acquired from microfluidic chemistry sets and genome sequencers in labs around the world at astonishing rates, and allows the environment — or the human body — to be monitored in real time by vast arrays of sensors. The fact that everyday computing is getting exponentially cheaper promises to vastly increase data flows of all sorts and to revolutionize the practice of science.

It is this remarkable growth that has allowed projects such as the LSST to be imagined — and which will surpass them before they are very old. It is not driven by science, but it has been of immense use to scientists, and will continue to be, if they can change the way science is done to make use of the great potential.

Scientists will increasingly have to rely on automation to extract useful knowledge from these vast data resources. As with computer-aided proofs in mathematics, such automation challenges the processes by which scientists gain insight and generate theories. What's more, science will increasingly be done directly in the database, finding relationships among existing data while someone (or something) else performs the primary collecting role. And this means that scientists will have to understand computer science in much the same way as they previously had to understand mathematics, as a basic tool with which to do their jobs.

Science can provide ideas and challenges that provoke the computer industry into moves it might not have made so quickly on its own.

But scientists can be more than just passive responders to change. Although the great trends in computing are driven by economic and technical forces external to the scientific world, science can provide ideas and challenges that provoke the computer industry into moves it might not have made so quickly on its own. The World Wide Web, after all, grew out of the needs of scientific data users. It was years after Tim Berners-Lee had put his vision of hypertext onto the Internet that it revealed its capacity to revolutionize fields from bookselling to campaign financing.

The computer industry knows that scientists can come up with strange ideas and requirements that may well, in time, have broader commercial application elsewhere. This is one of the reasons why Microsoft is engaging the scientific community with its new ‘Towards 2020 Science’ report on computers in science (see That report inspired this week's focus on computing in Nature. Microsoft is sponsoring free web access to our articles on the subject, although, as always, the content is exclusively Nature's responsibility.

As computing gets ever cheaper, quicker and more powerful, scientists would do well to remember that, by being a demanding and stimulating ‘user community’ that engages the interest of companies such as Microsoft, Google and Intel, they can influence the development of the field, to everybody's benefit.