Technology: Smartphone science

Journal name:
Nature
Volume:
531,
Pages:
669–671
Date published:
DOI:
doi:10.1038/nj7596-669a
Published online

Researchers are learning how to convert devices into global laboratories.

A decade ago, Dutch astronomer Frans Snik invented a simple optical device to measure the density of dust, soot and other particles, or aerosols, in the atmosphere that affect human health and the climate. He hoped to launch it into orbit around Earth aboard a satellite. But one afternoon in 2011, Snik held up a demonstration version of his device to an iPhone camera. The smartphone's screen displayed a rainbow of colours: Snik's optical device was converting incoming light into a spectrum that contained polarization information and channelling it into the camera. Snik realized that he could pair smartphones with the optical device and make the same kind of measurements that he and his colleagues planned to record from space.

An idea was born. “We thought, why not make use of a technology that millions of people carry around in their pockets anyway?”

iSPEX Project

Researchers measure atmospheric aerosols with iSPEX optical devices and smartphones.

By 2013, Snik and his colleagues at Leiden University in the Netherlands had given or sold a version of the optical device — called iSPEX — to more than 8,000 willing iPhone users across the country. The users followed instructions provided by an associated app to attach the optical devices to their iPhone cameras and photographed the sky in their local areas. Within a day, reams of crowdsourced spectra had stacked up in an online database, ready for analysis. It resulted in a Netherlands-wide map of atmospheric particles with unprecedented resolution (F. Snik et al. Geophys. Res. Lett. 41, 73517358; 2014) — several years before the proposed satellite launch and for a fraction of the original estimated cost. The team has since received funding from the European Union to repeat the project in 11 European cities.

Citizen science

Many researchers are finding ways to exploit smartphones. Snik's project, and those of some geophysicists, astronomers and other scientists who need huge data sets, go one step further. They recruit citizen scientists who use their own smartphones to collect data that would be difficult — if not impossible — to obtain in conventional ways. The various internal sensors that smartphones carry, such as cameras, microphones, accelerometers and pressure gauges, coupled with user-friendly apps offer a way for the public to contribute high-quality data. “There are tons of possibilities for science,” says Travis Desell, a computer scientist at the University of North Dakota who designs research projects that run on smartphones.

Scientists who want to exploit the potential of smartphones first need to assess whether the devices can obtain the measurements they need. They must then decide which software platform will optimize the proposed use, before ironing out any errors or 'bugs' in the apps that will be used to collect data. Scientists should also determine how to screen out invalid data. And they need to find ways to recruit participants.

Although recruiting the public isn't complicated, thanks to social media, it can still be time-consuming. Snik and his colleagues had a head start — the iSPEX project was covered by the Dutch media, which prompted a few thousand citizens to send in requests to participate. The team drummed up a similar number of contributors by collaborating with the charity Lung Foundation Netherlands in Amersfoort, which invited participation from supporters who were concerned about the effects of aerosols on health. Even so, the iSPEX researchers had to spend a year and a half on their crowdsourcing campaign, which involved uploading instruction manuals and video tutorials to a website, posting calls for support on social media and in online publications, and answering questions. But their efforts paid off when they received more than 6,000 submissions of data.

The more technical aspects of crowdsourcing data can be trickier to master, and it helps to have some technological savvy. Scientists will find it useful to know how to write an app or how to manufacture an inexpensive hardware 'add-on' (see 'How to create a hand-held research toolkit'). But if a researcher is not an adept programmer, help is available. Snik and his team turned to DDQ, a Netherlands-based company that creates apps that are tailored to citizen science. Researchers who lack funds for third-party support can learn to write an app themselves, thanks to a wealth of free online tutorials and discussion forums.

Box 1: How to create a hand-held research toolkit

Claire Welsh/Nature

Converting smartphones into a tool for citizen science is likely to require an app, and could also involve designing and manufacturing hardware accessories, or 'add-ons'. There is plenty of help available online for researchers who want to write their own app. Google's Android, the world's most popular mobile-device operating system, is supported by a development community that provides walk-through tutorials and guides to achieving specific functions.

The second-most popular mobile operating system, Apple's iOS, has a similar community called the Apple Developer Forums. And app developers for the Windows Phone can check in to the Microsoft Virtual Academy. Other websites offer free training, such as Alison, which takes novices through the app-writing process from start to finish.

The potential of apps for research has been recognized by Apple, which last year launched an online resource called ResearchKit. Developed with help from various universities and other research centres worldwide, ResearchKit is a set of tools and services that assist researchers to design and administer app-based studies for the iPhone (see Nature 531, 422423; 2016). The only catch is that ResearchKit is geared towards medicine. Apps that already benefit from it include mPower, which monitors people with Parkinson's disease, and GlucoSuccess, which assesses how daily activities affect glucose levels.

Whatever the research goal, it is best to start with the basics. “I read a few tutorials and wrote a small do-nothing app,” says particle physicist Daniel Whiteson at the University of California, Irvine, who is working on an app to record cosmic-ray events. “Then I slowly added functionality — such as accessing the camera and uploading data to another computer — until it was doing what I wanted.”

Creating add-ons for smartphones is a different challenge, but such devices need not be complex. In 2014, Steve Lee and his colleagues at the Australian National University in Canberra discovered that a smartphone camera could be given a magnification factor of up to 160 by attaching a single pea-sized lens. Costing less than an Australian cent (under US$0.01), the lens is created by allowing a droplet of polymer to harden on a curved substrate. “It forms a basic low-powered microscope system,” says Lee.

Lee's droplet lens builds on an existing smartphone instrument — its camera — but not all add-ons do. SCiO is a standalone near-infrared spectrometer developed by the start-up firm Consumer Physics in Israel. Due for release in July, the device will scan materials to provide molecular information and connect to a smartphone through Bluetooth wireless technology.

Researchers also need to decide which software platform to select. Snik and his colleagues chose the popular Apple iOS: the physical similarity between iPhone models made it easier to design a compatible add-on. But the leading platform, Google's Android, has advantages, too. It is less strict about the nature of apps and presents fewer barriers to its instrumentation.

Remote-sensing scientist Liam Gumley at the University of Wisconsin–Madison has developed an app that aims to improve weather forecasting by comparing photos of the sky taken from smartphones with satellite imagery. He has advice for anyone who is interested in smartphone-aided science: “Just do it!” Gumley recommends drawing up a set of storyboards that describe exactly what the app will do, what each screen will look like and what will happen when the user touches an onscreen control or a button. It is also a good idea, he says, to determine whether any data processing will be performed by the app or by a server online. Depending on the type of processing that is required, one might be faster than the other.

Big data

Researchers must also be ready with a database that can accommodate a deluge of data. “If you release the app globally, you may get more data than you expect within days,” warns Qingkai Kong, a PhD student in seismology at the University of California, Berkeley, who is working on MyShake, a seismology app. After extrapolating from a small group of users how much data he and his colleagues were likely to receive, they turned to Amazon Web Services to host their database. Other available cloud-computing services include the Google Cloud Platform and Microsoft Azure.

Once the data have been collected, it can be difficult to know whether they are reliable. Kong and his colleagues are refining MyShake so that it can distinguish between an actual seismic event and when a user is shaking a phone. A similar app, known as CSN-Droid and designed by scientists at the California Institute of Technology (Caltech) in Pasadena, was discontinued because it could not reliably make such distinctions. But Kong thinks that rigorous testing will reveal ways to improve MyShake's accuracy.

“Smartphones are very powerful and very flexible.”

Particle physicist Daniel Whiteson of the University of California, Irvine, is also tackling data reliability. He and his colleagues have developed an app called CRAYFIS (Cosmic Rays Found in Smartphones) that enables smartphone users to observe and record the particle debris that is generated when high-energy cosmic rays strike Earth's atmosphere (D. Whiteson et al. Preprint at http://arxiv.org/abs/1410.2895; 2014). If several hundred smartphones in a kilometre radius simultaneously detect a signal, or 'blip', the app registers the event as a cosmic-ray shower. The more blips that occur in a given radius, the greater the energy of the primary cosmic ray. But there is still the possibility that synchronous blips could originate from sources other than cosmic rays — including detector noise or ambient light.

Whiteson and his team hope to rule this out by recording the metadata that accompany blips, such as their time and location. If a smartphone is left in one place to record data, the researchers will be able to characterize sources of ambient light and noise so that genuine cosmic-ray signals become readily apparent. More than 150,000 people worldwide have already signed up to participate in the CRAYFIS study, but before they release the app officially, the researchers want to make sure it is free of performance issues that could drive contributors away. The team is currently running a test version of the app on 1,000 phones worldwide.

Despite the glitches, apps that crowdsource data are especially attractive for researchers because they can overcome issues that might prevent the collection of data. “The prospect that seismic data in large earthquakes can be obtained from consumer electronics is potentially transformative,” says Tom Heaton, a seismologist at Caltech. “One major obstacle to acquiring seismic data in a building is that the building owners are frightened by the prospect that researchers will uncover a critical safety issue.”

Just as smartphones have become indispensable for many scientists' day-to-day lives, they might also prove to be transformative vehicles for some experiments. “Gone are the days when governments would invest US$10 billion to $15 billion on new types of infrastructure, so it's important to think about the infrastructure that's already been built,” Whiteson says. “Smartphones are very powerful and very flexible. It's an enormous platform that we're only now beginning to think about for science.”

Author information

Affiliations

  1. Jon Cartwright is a freelance journalist in Bristol, UK.

Author details

Additional data