Abstract
Snowfall information at the scale of individual particles is rare, difficult to gather, but fundamental for a better understanding of solid precipitation microphysics. In this article we present a dataset (with dedicated software) of in-situ measurements of snow particles in free fall. The dataset includes gray-scale (255 shades) images of snowflakes, co-located surface environmental measurements, a large number of geometrical and textural snowflake descriptors as well as the output of previously published retrieval algorithms. These include: hydrometeor classification, riming degree estimation, identification of melting particles, discrimination of wind-blown snow, as well as estimates of snow particle mass and volume. The measurements were collected in various locations of the Alps, Antarctica and Korea for a total of 2’555’091 snowflake images (or 851’697 image triplets). As the instrument used for data collection was a Multi-Angle Snowflake Camera (MASC), the dataset is named MASCDB. Given the large amount of snowflake images and associated descriptors, MASCDB can be exploited also by the computer vision community for the training and benchmarking of image processing systems.
Measurement(s) | fall speed • snowfall images • snow crystals • geometrical characteristics • textural characteristics • snowfall microphysics |
Technology Type(s) | multi angle snowflake camera |
Similar content being viewed by others
Background & Summary
Snowfall is a phenomenon of extraordinary complexity across all temporal and spatial scales. At the scale of individual particles, the characteristics of a snow crystal or a snowflake (an aggregate of more individual crystals) depend in the first place on the environmental conditions upon formation and then on the environments encountered as the particle crosses the atmosphere from cloud top to ground deposition. Individual crystals may branch with each other and form aggregates1,2. They may collide with supercooled liquid water droplets that will freeze on them upon impact (riming3,4), they may sublimate or generate secondary ice by breaking up or splintering5. Measuring, describing and understanding the interactions occurring at the microphysical scale are complex tasks and important limiting factors to represent these mechanisms in numerical weather prediction models or climate models6. Because of computational constraints, these models are often unable to simulate processes at this scale and unavoidably adopt empirical parametrizations to get close to the observations7.
The research community compiled in the last half century a large inventory of snow crystal types8. This has been possible thanks to a double effort: years of manually-collected field observations from one side complemented by experiments in controlled laboratory conditions. Laboratory experiments have the special merit to quantitatively define the environmental factors leading to the formation of crystals with different shapes and types9. Although a taxonomy of individual crystals is nowadays established, much more has to be investigated about mass, density, fall speed and orientation and all the dynamical processes occurring while solid hydrometeors fall. A helping hand to the research in this field comes from the rapid development of snow particle imagers, either installed on aircraft to sample clouds and precipitation aloft10,11,12 or deployed at the ground level13,14,15,16. Ground-based imagers observe snowfall just before deposition and can provide information also on the fall speed, orientation and textural characteristics of falling snowflakes. The data collected with devices producing actual high resolution hydrometeor pictures, like the Multi-Angle Snowflake Camera (MASC)15, allow for immediate visual interpretation of the images. MASC data, for example, were used to perform hydrometeor classification17,18,19,20 as well as riming degree estimation of individual particles17 or, at the scale of a population of particles, to discriminate between snowfall and wind-blown snow21.
Three techniques have appeared in the literature to classify MASC images into solid hydrometeor types. In a supervised learning setting, Praz et al.17 employed expert knowledge for feature extraction and multinomial logistic regression for snowflake classification, while Hicks et al., 2019 and Key et al.18,20 described the use of convolutional neural networks architecture for the same purpose. On the other hand, Leinonen et al.19 presented an unsupervised approach where the expressive power of generative adversarial networks is employed to extract snowflake image characteristics without human expertise, which are then used to perform clustering using k-medoids.
MASC data have been exploited to propose three dimensional reconstruction techniques22,23 and have been more in general proven useful to better describe, quantify and interpret the microphysical properties of falling snow3,24,25,26,27.
Although the potential of MASC is indubitable, the studies listed above focused on individual field campaigns and/or on the development of specific retrievals possibly obtained using different data filtering and preprocessing techniques. This work aims to fill such gap and to provide the scientific community with a standardized large database (named MASCDB) covering at the present time ten field campaigns worldwide. The database includes MASC images, image descriptors and precomputed output of existing classification and retrieval algorithms. Raw MASC images have been treated within a homogeneous processing chain and additional effort has been devoted to add basic environmental information to all the collected measurements. Additionally, an intuitive and well documented python package has been developed to facilitate MASCDB data extraction, manipulation, exploration, visualization and to provide the possibility to extend the database with additional image descriptors and retrievals.
MASCDB is the first outcome of an international collaborative effort to promote exchange of data across the precipitation research scientific community and this precompiled dataset can be useful in several ways. First, it will allow to enhance scientists’ productivity, by focusing directly and efficiently on the improvement/development of microphysical retrievals and the analysis of snowflake characteristics, without the need to devote time to the reimplementation of preprocessing and retrieval algorithms which are already provided by this study. The large number of snowflakes available will allow to better characterize snowflake variability, while the included snowflake property estimates unlock potential to advance the accuracy of weather radar snowfall microphysical retrievals, and the robustness of microphysical schemes in weather and climate models.
The complexity of snowflake shapes in MASCDB images and their translation- and rotation-invariant semantic characteristics represents an ideal test-bed for the development and improvement of representation learning and image classification algorithms. Therefore, MASCDB also holds potential to be adopted as a benchmark dataset by the computer vision and machine learning community for the development of classifiers, as well as to test the ability of representation learning and generative algorithms to disentangle, extract, and reproduce the complex variety of snowflakes that nature can offer.
Finally, thanks to a well documented support software and to the overall compact size of the database, MASCDB could emerge as a useful tool for educational activities and exercise sessions in atmospheric and computer vision curricula.
MASCDB can be accessed on Zenodo28, while the pymascdb package at https://github.com/ltelab/pymascdb.
Methods
Multi-angle snowflake camera (MASC)
The Multi-Angle Snowflake Camera (MASC15) is an instrument able to collect a triplet of high resolution images of falling hydrometeors and to provide an estimate of their fall speed. A MASC, as illustrated in Fig. 1, is composed of three high resolution co-planar cameras with the same focal point situated at an approximate distance of 10 cm. The cameras are separated by 36° with respect to the next one such that a picture of each snowflake can be obtained simultaneously at an angle of 0° and ±36°, covering thus a total span of 72°. Two infrared (IR) emitter-detector pairs are triggering the cameras and three LED spotlights illuminate the targets. The IR arrays are separated vertically by 32 mm, providing in this way an estimate of particle fall velocity. The data processed and provided here have been collected with two identical MASC system, equipped with three 2448 × 2048 pix cameras of 12.5 mm focal length (f) and aperture f/5.6 operating with a maximum acquisition rate of about 2 Hz and an exposure time of 40 μs17. The pixel resolution in this set up is 33.5 μm. The sampling area is relatively small (nominally, according to specifications is 2.5 cm2) and this helps to avoid as much as possible multiple triggers of the IR detector by different particles. In calm air this is generally true, but as soon as the wind speed increases the sampling of larger snowflakes becomes less efficient29 and at the same time multiple triggers from smaller snowflakes or blowing snow particles can occur. In case of wind-blown snow, which is snow recirculated from the ground, several small particles may be present in the measurement area at the same time21. This is a known limitation and one of the reasons behind our effort to provide environmental information within our database, as detailed in the following sections.
Data processing chain
Raw data consists of a triplet of 2448 × 2048 gray-scale images (see Fig. 2) with values in the range 0–255 (black-white) collected by the three cameras. As a reference, CAM0 is the leftmost camera, for an observer facing the instrument (Fig. 1). The images are at first cropped around the borders. All images are cropped at the top and at the bottom, in order to remove the image background not belonging to the measurement area. CAM0 (CAM1) images are further left- (right-) cropped as the bright LEDs of the infrared beams appear in the pictures. To reduce the background noise, pixels with values lower than 12 are set to 0.
After this preliminary treatment, the images are processed in order to identify connected regions of interest (ROIs). To this end, the greyscale images are binarized. Any hole (single 0 pixels completely surrounded by ones) in the resulting mask is filled with ones. The mask is applied to the original gray-scale image and it is used to crop and extract around every separate ROI. ROIs with an area smaller than 10 pixels are discarded as well as ROIs touching the borders of the image.
The ideal measurement concept is the following: an individual snowflake crosses the measurement area, triggers the two IR detectors, a triplet of pictures is collected and a unique ROI is found on each image. In reality, in our dataset, individual ROIs are found in an image only approximately once every eight triggers. When precipitation is intense or mixed with blowing snow, or a large number of small particles are present, multiple objects appear on each image. Consequentially, the next step involves the matching of ROIs of individual cameras into triplets that are associated to the same snowflake. This is done by comparing their average vertical position and average vertical size, as this is a common dimension among the three cameras and it should approximately match. For each ROI, the central camera is taken as a reference and matches are searched in the other cameras. A best match is found among the particles having a maximum difference in vertical position and dimension with respect to the reference that is within 50 pixels (or within 100% of the reference value if this is larger than 50 pixels). Among all the matching triplets, we decided to keep only the one that is on average the sharpest and brightest, assuming that it corresponds to the best placed particle within the sampling area. The parameter used as quality discriminant is named AF (Area-Focus):
where A is the area of the ROI in pixels and F is a focus factor defined as:
with \(\bar{I}\) the mean intensity of pixels belonging to the ROI and \(\bar{{R}_{I}}\) the mean range intensity (local 3 × 3 max–min operator) of the same pixels.
The available data are further filtered prior inclusion into MASCDB in order to eliminate images with poor quality. The quality parameter ξ (as thoroughly described in Appendix B of17) is used as a discriminant. ξ quantifies the blurriness of the images as much as possible independently from their brightness. ξ values above 9 usually correspond to very sharp images, while values below 8 to very blurry ones. In previous studies, thresholds on ξ of 917 or 1019 have been applied for the purpose of hydrometeor classification, which requires very sharp and high-quality images promptly recognizable by visual inspection. Some geometrical descriptors, less influenced by texture, may still be quantitatively valuable even for low-quality images and therefore we selected a more permissive threshold for our database. A triplet of ROIs is accepted if each view has ξ of at least 8 or alternatively if at least one view has ξ larger than 8.5. The last condition in particular serves a double purpose: it preserves individual good quality images in overall bad quality triplets and it allows to include in the database also low quality images potentially useful for future image classification applications. As summarized in Table 1, MASCDB currently includes 851’697 triplets, corresponding to 2’555’091 snowflake images.
As it will be detailed in the following section (Data Records), MASCDB includes the actual gray-scale images of the ROIs of each triplet, in addition to a set of precomputed geometrical and textural parameters as well as a set of precomputed retrievals from published algorithms (Supplementary Tables 3 and 4). The geometrical and textural descriptors of each ROI are in large part the same ones as derived in Praz et al.17. A few visual examples are depicted in Fig. 3 and a complete list with relevant references and description is summarized in Supplementary Table 4.
Data from retrieval algorithms
The precomputed products provided in MASCDB and listed in the entries of Supplementary Tables 3 and 4 are briefly describe here.
Fall speed
The fall speed of the hydrometeors observed by the MASC is estimated by the difference in time between the trigger of the upper and lower IR emitter-detector pairs. This attribute is available in MASCDB with name flake_fallspeed. It has been shown in past research that wind and turbulence affect this measurement3,29; these studies suggest that, in order to maximize the reliability of the estimate, the MASC should be sheltered from the wind with a double fence and wind speed should be below 5 ms−1. As a reference, in MASCDB, about 35% of the data are associated to winds below this threshold. If we consider only the two field campaigns during which a double fence was available, this percentage drops to 11.5%.
Hydrometeor classification
The output class of the hydrometeor classification of Praz et al.17 is provided together with the classification probability of the assigned class. The output is provided both individually for each ROI in a triplet and as a unique value considering the triplet images altogether. The classification method discriminates between: small ice particles, aggregates, planar crystals (sectored plates, plates, stellar crystals/dendrites), columnar crystals (needles, columns), graupel and combination of planar crystals and columns (for example capped columns). It must be noted that the method does not discriminate some important ice-phase hydrometeors like hail, bullets (probably classified as columnar crystals) or combination of bullets as bullet rosettes. Hydrometeor classification outputs are stored in MASCDB in attributes starting with snowflake_class_. An example of good quality and high probability image triplets in the database belonging to the six hydrometeor classes is illustrated in Fig. 4. These correspond to good cases where the overall classification probability is very high: the hydrometeor types are clearly recognizable in every image of the triplet. It can also happen that only one or two cameras provide high-quality classification or, because of the different viewpoints, the classification output differs among the cameras (a typical example is a planar crystal that appears as a needle if observed from certain views). This is the reason why we stored information about the microphysical retrievals both for individual cameras and for the aggregated triplet.
It must be noted that the method of17 was originally developed from a training set of data belonging to only 2 out of the 10 field campaigns we provide in MASCDB. This opens the possibility for future researchers to exploit the additional data and fine-tune the classification, for example by discriminating between different types of planar crystals as plates and dendrites.
Riming degree estimation
The estimation of the riming degree of the particles, also described in Praz et al.17, is provided in an analogous way both for individual cameras and aggregated for the triplet. The method discriminates between five discrete riming degree levels (unrimed, rimed, densely-rimed, graupel-like, graupel) and also provides a continuously varying riming degree Rc defined in the interval 0–1, with 1 being graupel. The relevant attributes in MASCDB start with riming_. Riming degree estimation presents some known limitations in the experience of the authors: small and bright particles may be classified as graupel and more in general their riming degree level can be unreliable. Similar caveats apply to out-of-focus images.
Melting snow identification
Liquid water droplets on the surface of snowflakes appear as very bright spots on the ROIs and therefore it is possible to try to identify melting snow visually. This is the third retrieval described in Praz et al.17. The continuous probability of melting in the interval 0–1 is provided as well as a discrete flag (attributes starting with melting_). The estimation is conducted, and separately stored, both on individual views (individual cameras) and as an average value of each triplet. When the probability of melting is larger than 0.5, the degree of riming is not anymore reliable and the riming degree information is set to undefined in the data.
Although we did not censor the output of hydrometeor classification, riming degree estimation and melting according to image quality ξ, we recommend to treat with care the estimates corresponding to ξ < 9 as they did not belong to the original training set used for algorithm development. For the sake of completeness, we also flagged with dedicated boolean attributes (attributes starting with hl_, meaning “human label”) the images which belonged to the visually-interpreted training set employed by Praz et al.17 in order to provide the readers with a benchmark to test eventual new and improved methods.
Mass, mass distribution and volume estimation
The recent 3DGAN (3-Dimensional Generative Adversarial Network) method described in Leinonen et al.23 showed some promising results concerning the estimation of mass and three-dimensional properties using as input a triplet of MASC images. 3DGAN is based on machine learning, using a large number of simulated snowflakes as training set. We provide, in the variables starting with gan3d_ the estimation of total mass, volume and radius of gyration for the snowflakes for which 3DGAN could provide an estimate. 3DGAN requires good quality images in all the three views and it is not suited for particles classified as small_particle or particles undergoing melting and for this reason not all the triplets in MASCDB fulfill these requirements. Furthermore, some limitations of 3DGAN must be mentioned here. The simulated snowflakes used to train the algorithm included aggregates of various monomers and of different degrees of riming, but they were generated at a resolution of 40 μm. This is comparable with the resolution of the MASC but not enough to fully depict the interaction of supercooled liquid water droplets of micron-size with the complex spatial structures of typical snowflakes. Additionally, 3D-GAN has been validated on a limited number of physical samples (3D printed snowflakes), also having a spatial resolution of 40 μm.
Blowing snow identification
Blowing snow can be an unwanted source of noise in the data collected by the MASC or constitute relevant information in itself30. During blowing snow events, the MASC is triggered more often and many small ROIs may be recorded in the raw images of the three cameras. The method described in Schaer et al.21 employed image processing techniques on the raw MASC images and estimated if the population of particles recorded by the MASC instrument at a given time is more likely to belong to one of three classes: pure blowing snow, pure precipitation or precipitation mixed with blowing snow. In the last case, a mixing index ranging from 0 to 1 is also provided. We provide the predictions of this methodology with variables starting with bs_.
The availability of this estimate is useful for instance as a filter, when measurements from different field campaigns are merged together. A field campaign like Davos-2015, conducted in a wind protected environment, has only 0.3% of total triplets estimated to belong to a pure blowing snow population while for a field installation like APRES3-2017 on the coasts of Antarctica where katabatic winds blow31 this proportion increases to 28%.
Environmental data
It is relevant and may be useful to have access to basic environmental conditions in the proximity of the MASC instrument at the time of each image record. MASCDB provides (with attributes starting with env_) information about air temperature, relative humidity (with respect to water), wind speed, wind direction and atmospheric pressure. During all the campaigns this information was available in the close proximity (within 10 m) of the instrument, although from different sensors and different temporal resolutions ranging from seconds to minutes, as summarized in Table 2. Some data were collected by research instruments, temporarily deployed during selected field campaigns (using in all cases a Vaïsala WXT520 weather station) while in other cases we could access data collected by operational weather services. Data have been re-sampled at a common interval of one minute before being matched temporally to the closest MASC observation. We are confident that variables as temperature, humidity or pressure are comparable among different field installations, while wind direction and wind speed should be treated with care when data of more field campaigns are aggregated, as the local set-up of the instruments (local wind shielding, etc) may play a significant role, especially for data not collected following operational protocols.
Field campaigns
We provide data collected during 10 field campaigns from 2015 until 2021, as listed in Table 1. The campaigns have been conducted in different locations worldwide. Even though the exact geographical locations and other relevant information is readily available in the data themselves, it is worth to briefly describe the context and specificity of each installation.
Davos-2015: 46.8297 N, 9.8093 E, 2540 m amsl
This installation took place in an optimal setting: the MASC instrument was set up in a site belonging to SPICE32 (Solid Precipitation Intercomparison Experiment) within a double fence wind protection, thus particularly protected from wind-related biases21,29.
APRES-2016 and APRES3-2017: -66.6628 N, 140.0014 E, 41 m amsl
During the international APRES3 program30,33 (Antarctic Precipitation: Remote Sensing From Surface and Space), a MASC instrument was deployed at the Antarctic French base Dumont d’Urville. Installed on the Antarctic coast, the MASC was often exposed to strong winds and consequently it frequently collected images of wind-blown snow. Given the remote location, the data collected here are particularly interesting and still largely unexploited. The second campaign in this location (APRES3-2017) includes unique data collected also during the months of Antarctic winter.
Valais-2016: 46.1222 N, 7.2122 E, 2370 m amsl
The instrument was installed in a ski resort in the Swiss Alps.
ICEPOP-2018: 37.6652 N, 128.6996 E, 789 m amsl
ICEPOP (International Collaborative Experiment for PyeongChang Olympic and Paralympics) was an international research effort conducted during the winter Olympics game in 2018 in Korea. The MASC instrument was installed, as for Davos-2015, within a double fence protection, in the measurement site of Mayhills34.
PLATO-2019: -68.5752 N, 77.9659 E, 10 m amsl
The measurements of this installation were collected in the Davis research station in Antarctica during a summer field campaign35. Only a limited number of triplets have been collected by the MASC instrument in this occasion.
Davos-2019: 46.8450 N, 9.8716 E, 1512 m amsl
MASC was installed in the Alps of eastern Switzerland in the context of Role of the research program: Aerosols and Clouds Enhanced by Topography on Snow (RACLETS)36). The installation site is not the same as in Davos-2015.
Jura-2019: 46.6702 N, 6.3125 E, 1045 m amsl
The instrument was deployed in western Switzerland, in the Jura mountains.
POPE-2020: -71.9499 N, 23.3471 E, 1382 m amsl
This installation took place in Antarctica, as a contribution to the Princess Elizabeth Antarctic Orographic Precipitation Experiment (POPE). The station was located in Queen Maud Land (East-Antarctica). Due to the presence of mountains of almost 3000 m of altitude in a 25 km radius from the station, the main goal of the campaign was to study the effects of the mountain range on clouds and snowfall at the site. The region, as well as the scientific base, are well described for example in Gorodetskaya et al.37.
ICEGENESIS-2021: 47.0830 N, 6.7922 E, 1018 m amsl
The instrument was installed in the Swiss Jura mountains in the framework of a field experiment contributing to the ICEGENESIS project (https://www.ice-genesis.eu/).
Data Records
MASCDB can be accessed on Zenodo28. This section provides information about the data format and the data records. Any user should be able to read the data with the programming language of its choice. We recommend nevertheless to take advantage of the pymascdb package provided and described in the Usage Notes section. The dataset is composed of four Apache parquet (https://parquet.apache.org/, last accessed on Nov 1st 2021) binary files containing scalar descriptors of individual snowflakes and one Zarr store where the gray-scale images of the corresponding ROIs are stored.
The four Apache parquet files are:
-
MASCdb_triplet.parquet: records listed in Supplementary Table 3.
-
MASCdb_CAM0.parquet: records listed in Supplementary Table 4.
-
MASCdb_CAM1.parquet: records listed in Supplementary Table 4.
-
MASCdb_CAM2.parquet: records listed in Supplementary Table 4.
The MASCdb_triplet.parquet file contains information that are valid for a triplet of images. For example, to cite a few entries, it includes environmental information, blowing snow detection, 3D-reconstruction parameters and MASC-estimated fall speed. The other three Apache parquet files (one for each camera) include the geometrical and textural descriptors as well as the retrievals computed over the ROIs identified in each camera view.
The Zarr store (named MASCdb.zarr) contains gray-scale images of all the triplets in the dataset. The four dimensional tensor with all triplet images has dimensions (number of triplets, y, x, camera id), shape (851697, 1024, 1024, 3) and is saved to disk in a Zarr array with chunk size (256, 1024, 1024, 3). The reason behind the choice of Zarr, with respect to other classical formats in use in the geoscience community for storing n-dimensional tensor, is that HDF and netCDF38 formats do not allow (without specialized installations) multi-threaded data input/output in python, introducing performance bottlenecks for the distributed computations enabled by the pymascdb introduced in the Usage Notes section. In order to work on common dimensions, every ROI of the triplet is centered into a grid of size 1024 × 1024 pixels by adding black (0) pixels around their boundary. Given a typical pixel size of 33.5 μm, this correspond to a square box of about 3.5 cm side. Only one snowflake in the entire database exceeded the size of this bounding box of a few pixels; we decided therefore to keep the size of 1024 pixels, as it is a convenient power of 2, rather than further increase it.
Technical Validation
Snowflake geometry and microphysics
The availability of several geometrical and textural descriptors of the images, in combination with the output of retrieval algorithms tailored on snowflake microphysical characteristics, allows MASCDB users to perform all types of multivariate analysis.
For example, Fig. 5 stratifies the distribution of two geometrical descriptors according to the hydrometeor classification output. While it is reassuring but not surprising that aggregate snowflakes reach larger sizes with respect to other snowfall types, it is very interesting to observe the behavior of a more “complex” parameter: sym_P6 (see Supplementary Table 4). The family of sym_Pn parameters are related to the n-fold symmetry of the images. The interesting signature is the increase of 6-fold symmetry for the hydrometeor category of planar crystals. Planar crystals include stellar and plate-like structures, as shown in Fig. 4 where the hexagonal basic structure of ice crystals is the most noticeable.
Data can be stratified also according to the riming degree level. Figure 6 provides the conditional distribution of the continuous riming degree index Rc (or riming_deg_level) with respect to the five discrete riming degree classes (left panel), as well as the distribution of a descriptor designed to capture particle compactness based on areal approximations with a fitted ellipse (compactness, right panel). While looking for example at compactness, a physically reasonable trend emerges: increasing riming degree levels are associated to more compact particles. The noteworthy exception is the large variability of compactness observed in particles classified as unrimed, so it is important to recall that: (i) compactness is only one among many descriptors used to depict the riming degree, (ii) the estimation of riming degree suffers of the limitations discussed previously, in case of small and bright particles.
Environmental conditions and MASC measurements
The availability of co-located environmental measurements is a simple, important, and not very often available, source of information for microphysical investigations.
Figure 7 shows the MASCDB relationships between the identification of melting particles17 and air temperature, as well as the prediction of blowing snow occurrence21 and wind speed, while Fig. 9 presents the distribution of air temperature for aggregates larger than 5 mm. The results are physically-sound and they cross-validate the estimation methods which, in both cases, do not employ any environmental information for class predictions but use only the appearance of individual (or of a population of) particles. Particles labeled as melting are mostly measured around a temperature of 0 °C and there is a clear signature of particles associated to blowing snow when wind speed is very high. The large values of wind speeds must not surprise the reader, recalling that some field campaigns have been conducted in Antarctica where strong katabatic winds blow31. It is worth to underline, regarding the conditional distribution of air temperature with respect to melting particles, that a more appropriate environmental variable to consider for investigating this phenomenon would be the wet bulb temperature, which can be estimated by using relative humidity and atmospheric pressure data provided by MASCDB.
Concerning the impact of winds on MASC measurements, previous studies have shown that wind speed affects the quality of hydrometeor fall speed data as recorded by the MASC24 and it reduces significantly the probability to observe particles of larger dimension29. A qualitative confirmation of this behavior of the MASC for the data in our database can be observed in Fig. 8: the larger the wind speed, the lower the maximum dimension of the observed particles.
Parameterization of microphysical properties
MASCDB additionally paves the way for an enhanced understanding of the microphysical properties of frozen hydrometeors. This pertains to the field of weather and climate science, both through cloud and precipitation models and the microphysical parameterizations they rely on, and for the remote sensing community in the development of more realistic scattering models.
For instance, the mass estimates from 3D-reconstruction of MASC triplets23 allow to compute mass-dimensional relations, which are typically represented with power laws of the form \(m=a{D}_{max}^{b}\). For each class of snow particle, each precipitation event in the database yields a pair (a, b), as shown in Fig. 10 for planar and columnar crystals, aggregates and graupel. These results show a relatively good agreement with results from previous studies39,40,41,42,43,44, in which (a, b) were derived from in-situ measurements. The largest discrepancy is seen for planar crystals; this could be related to the more challenging identification of this particle type, whose aspect can be quite different in the different camera views. An interesting outcome of this analysis is the strong correlation that is observed between exponent and prefactor values, which supports the findings of45, and which highlights that snow particle growth is driven by highly structured mechanisms. Additionally, it appears that while the different particle types have relatively similar a-b relations, the distribution of the exponent b varies significantly; in particular, graupel particles have a fractal dimension close to 3, consistent with their sphere-like geometry.
Besides evidencing these features, this method provides distributions for a and b parameters, rather than point values as were typically computed from previous studies. This in turn allows for a statistical approach in the microphysical representation of frozen hydrometeors, which could additionally be fine-tuned according to riming degree, environmental conditions, campaign location, etc. Similar results are obtained when other properties are studied such as the area-size relation, as well as aspect ratio distribution, and potentially multiple other features.
Perspectives on snowflakes representation learning, classification and generative modeling
As demonstrated by recent studies18,19,20,23, the usage of specialized deep learning architectures offers an interesting perspective for improvement and automation of MASC data information extraction. While supervised classification methods have been developed for this purpose in recent years18,20, their ability to generalize is limited by the representativeness of their labeled training datasets and the manual effort required for its assembly. Additionally, the definition of classes is subject to the scientist judgment and introduces artificial class boundaries which might not reflect the actual formation and continuous evolution of snowflakes shapes occurring in nature.
An alternative approach, which Leinonen et al.19 have started to explore, consists in the development of unsupervised techniques to divide a dataset into classes without expert-knowledge (training labels) provided by the scientists. Such methods rely on the ability of representation learning algorithms46,47,48,49,50,51,52,53,54 to recognize the important modes of variations in the images and encode their disentangled representation into a (latent) vector. Most of this continuous and discrete generative factors are usually constrained to be invariant to translation, rotation, and orientation of the image. Depending on the dimensionality of the latent space, these image descriptor factors can be further reduced in dimensionality55,56,57 to better investigate the encoded properties in low-dimensional manifolds58,59,60. Unsupervised classification can be performed by applying custom clustering algorithms on the latent factors61,62,63 or by including clustering specific loss into the neural networks64,65,66. Figure 11 illustrates the representational disentangling ability of such algorithms, by displaying MASC images randomly sampled for each cluster obtained by applying self-organizing maps67 to the factors of variations derived by the algorithm described in Leinonen et al.19.
The latent factors of a specific MASC image can be exploited also to research similar ones68,69: either by searching for existing MASC images whose latent factors lie closely to those of the anchor image, or alternatively by sampling from the latent space joint distribution and exploit the algorithm decoder/generator to simulate MASC images associated to the encoded latent structure. Such workflows would allow, for example, to investigate which environmental conditions are associated to similar snowflake structures, or alternatively to tentatively simulate the smooth evolution of a snowflakes conditional to some latent factor of variations by traversing the latent space.
Advances in generative adversarial networks70,71,72,73,74 have also improved the algorithmic capabilities for image enhancement75 and restoration76,77,78,79, which could be exploited to improve the quality of MASC images flagged to be of low quality (low ξ) and consequently improve the number of usable MASC images. Similarly, although algorithms for snowflake 3D reconstructions are already available22,23, it is foreseeable that advances in 3D modeling80 and representation learning81 will allow to further refine snowflake 3D reconstructions, microphysical retrievals and potentially scattering simulations.
Usage Notes
MASCDB data are shared with the scientific community together with a python82 package named pymascdb which aims to facilitate data manipulation, analysis and visualization, besides providing utility methods for future database extension and algorithms development.
The pymascdb python package documentation, available at https://github.com/ltelab/pymascdb and https://pymascdb.readthedocs.io/en/latest/index.html, describes the usage of all the package functions and provides several tutorials to introduce the users to the data structure and the MASC_DB class methods.
When using pymascdb, all MASCDB data can be read in python into a MASC_DB class instance. The tabular data stored into the Apache parquet files are loaded in RAM memory into pandas83 objects, while the four-dimensional tensor with MASC images saved in the Zarr store is opened in “lazy mode” using Xarray84 with Dask85 as the array back-end. This means that the MASC images are not loaded into RAM, which allows to stream computation also on computing systems with not enough RAM to load the entire four-dimensional array into memory. Since operations on Dask arrays are lazy, operations such as filtering, sorting and image enhancement queue up as a series of sequential tasks mapped over blocks of data (Dask chunks). No computation is performed until the actual values of specific data chunks need to be computed (i.e. for visualization or image features extraction). At that point, data is loaded into memory and computation proceeds in a streaming fashion, block-by-block (Dask chunk by Dask chunk). More information can be found in the documentation of Xarray and Dask python packages.
Once a MASC_DB class instance is created, pymascdb enables the users to benefit from implemented class methods specifically designed to ease data manipulation operations such as data subsetting, filtering, sorting and reordering. For example, data can be filtered:
-
by campaign (select_campaign, discard_campaign)
-
by snowflake class (select_snowflake_class)
-
by riming degree (select_riming_class)
-
by blowing snow occurrence (select_precip_class)
-
by apparent melting features (select_melting_class)
or reordered according to a specific image descriptors of a camera, for instance as:
mascdb.arrange (‘cam0.Dmax’, decreasing = True)
Many more examples are provided in the tutorials accompanying pymascdb.
To empower exploratory data analysis of MASCDB columnar format data (i.e. triplet, cam0, cam1, and cam2), pymascdb implements an accessor to the Seaborn86 data visualization library, to facilitate the visualization of multivariate relationships in multifaceted ways (see Figs. 5–9). Similarly, several plotting methods, built upon Xarray and Matplotlib87 libraries, are implemented to instantaneously display multiple MASC triplet images as well as camera-specific snowflakes. In order to improve the visual appearance of MASC images, pymascdb provides the possibility to correct the “raw” images with contrast stretching as well as global, adaptive and local histogram equalization algorithms. The plotting methods also allow to zoom close to the bounding box of the MASC image. For example, Fig. 4 was created with the pymascdb method plot_triplets, correcting the images with adaptive histogram equalization and zoom enabled.
To promote more detailed investigations, pymascdb also offers a utility to customize the definition of snowfall events. While the default pymascdb settings assumes a snowfall event to be defined by at least one MASC image acquisition within a time interval of 4 hours, the user can redefine the events by specifying parameters such as:
-
maximum_interval_without_images
-
event minimum_duration or
-
event minimum_n_triplets.
While changing maximum_interval_without_images changes only the number of total events, the specification of the latter two parameters trigger a filtering of the dataset. After redefining snowfall events, the user can narrow down their selection of relevant events through implemented methods, for instance:
-
select short events with mascdb.select_events_shortest(n = 20) or
-
long-lasting snowfall mascdb.select_events_with_duration(min = np.timedelta64(2, ‘D’)).
MASCDB summary statistics of events and campaigns are also available typing mascdb.event and mascdb.campaign.
In conclusion, in the perspective of future inclusion of additional MASC images by upcoming campaigns and other institutions, pymascdb also implements tools to update or modify the MASCDB dataset presented in this manuscript, as well as it implements routines that might facilitate the development of new retrieval algorithms by the scientific community. For example,the method compute_2Dimage_descriptors enables the scientist to focus only on the development of a single custom-designed function that extract descriptors from a single image, and then pymascdb performs the heavy lift to scale the computation over all MASCDB images exploiting all the cores or cluster nodes available. The usage of a Dask array as Xarray back-end enables the user to choose between all available Dask schedulers which allows to perform multithreaded, multiprocess or distributed computing. If the Dask distributed scheduler is used (the current default), the image extraction computations can be profiled and monitored interactively in the Dask Dashboard.
A basic example of syntax is the following one, showing how Fig. 9 was generated, with all the filters applied to the data: #--------------------------------------------------------
# Import MASC_DB instance from api.py
from mascdb. api import MASC_DB
# Import plotting utils
import matplotlib as mpl
import matplotlib.pyplot as plt
# local path where all MASCDB files must be
dir_path = “/data/MASC_DB/”
# Create MASC_DB instance
mascdb = MASC_DB(dir_path = dir_path)
# Select only precip (discard mixed and blowing snow) and only aggregates
mascdb = mascd.select_snowflake_class(‘aggregate’). select_precip_class(‘precip’)
# Select only particles larger than 5 mm on Dmax
idx = mascdb.triplet[‘flake_Dmax’] >0.005
mascdb = mascdb.isel(idx)
# Plot taking advantage of seaborn integration
plt.figure(figsize = (10,6))
ax = mascdb.triplet.sns.histplot(x = ‘env_T’, kde = True, stat = ‘probability’, common_norm = False, binwidth = 0.5)
plt.show()
#------------------------------------------------------------
Code availability
The pymascdb package to manipulate the data and described in the previous section can be freely accessed at https://github.com/ltelab/pymascdb. Code documentation, examples and installation instructions are also available at https://pymascdb.readthedocs.io/en/latest/index.html.
Change history
23 May 2022
A Correction to this paper has been published: https://doi.org/10.1038/s41597-022-01375-6
References
Lawson, R. P., Stewart, R. E. & Angus, L. J. Observations and Numerical Simulations of the Origin and Development of Very Large Snowflakes. Journal of Atmospheric Sciences 55, 21 (1998).
Karrer, M. et al. Ice Particle Properties Inferred From Aggregation Modelling. Journal of Advances in Modeling Earth Systems 12, https://doi.org/10.1029/2020MS002066 (2020).
Garrett, T. J. & Yuter, S. E. Observed influence of riming, temperature, and turbulence on the fallspeed of solid precipitation. Geophysical Research Letters 41, 6515–6522, https://doi.org/10.1029/2020MS00206610.1002/2014GL061016 (2014).
Enzmann, F. et al. 3-D imaging and quantification of graupel porosity by synchrotron-based micro-tomography. Atmospheric Measurement Techniques 4, 2225–2234, https://doi.org/10.1029/2020MS00206610.5194/amt-4-2225-2011 (2011).
Korolev, A. & Leisner, T. Review of experimental studies of secondary ice production. Atmospheric Chemistry and Physics 20, 11767–11797, https://doi.org/10.1029/2020MS00206610.5194/acp-20-11767-2020 (2020).
Morrison, H. et al. Confronting the Challenge of Modeling Cloud and Precipitation Microphysics. Journal of Advances in Modeling Earth Systems 68, https://doi.org/10.1029/2020MS00206610.1029/2019MS001689 (2020).
Tapiador, F. J., Sánchez, J.-L. & García-Ortega, E. Empirical values and assumptions in the microphysics of numerical models. Atmospheric Research 215, 214–238, https://doi.org/10.1016/j.atmosres.2018.09.010 (2019).
Kikuchi, K., Kameda, T., Higuchi, K. & Yamashita, A. A global classification of snow crystals, ice crystals, and solid precipitation based on observations from middle latitudes to polar regions. Atmospheric Research 132-133, 460–472, https://doi.org/10.1016/j.atmosres.2013.06.006 (2013).
Libbrecht, K. G. The physics of snow crystals. Reports on Progress in Physics 68, 855–895, https://doi.org/10.1088/0034-4885/68/4/R03 (2005).
Heymsfield, A. J., Bansemer, A., Schmitt, C., Twohy, C. & Poellot, M. R. Effective Ice Particle Densities Derived from Aircraft Data. Journal of Atmospheric Sciences 61, 22 (2004).
Jackson, R. C. & McFarquhar, G. M. An Assessment of the Impact of Antishattering Tips and Artifact Removal Techniques on Bulk Cloud Ice Microphysical and Optical Properties Measured by the 2D Cloud Probe. Journal of Atmospheric and Oceanic Technology 31, 2131–2144, https://doi.org/10.1175/JTECH-D-14-00018.1 (2014).
Praz, C., Ding, S., McFarquhar, G. & Berne, A. A Versatile Method for Ice Particle Habit Classification Using Airborne Imaging Probe Data. Journal of Geophysical Research: Atmospheres 123, https://doi.org/10.1029/2018JD029163 (2018).
Kruger, A. & Krajewski, W. F. Two-Dimensional Video Disdrometer: A Description. Journal of Atmospheric and Oceanic Technology 19, 16, https://doi.org/10.1175/1520-0426(2002)019 (2002).
Newman, A. J., Kucera, P. A. & Bliven, L. F. Presenting the Snowflake Video Imager (SVI). Journal of Atmospheric and Oceanic Technology 26, 167–179, https://doi.org/10.1175/2008JTECHA1148.1 (2009).
Garrett, T. J., Fallgatter, C., Shkurko, K. & Howlett, D. Fall speed measurement and high-resolution multi-angle photography of hydrometeors in free fall. Atmospheric Measurement Techniques 5, 2625–2633, https://doi.org/10.5194/amt-5-2625-2012 (2012).
Kuhn, T. & Vázquez-Martín, S. Microphysical properties and fall speed measurements of snow ice crystals using the Dual Ice Crystal Imager (D-ICI). Atmospheric Measurement Techniques 13, 1273–1285, https://doi.org/10.5194/amt-13-1273-2020 (2020).
Praz, C., Roulet, Y.-A. & Berne, A. Solid hydrometeor classification and riming degree estimation from pictures collected with a Multi-Angle Snowflake Camera. Atmospheric Measurement Techniques 10, 1335–1357, https://doi.org/10.5194/amt-10-1335-2017 (2017).
Hicks, A. & Notaroš, B. M. Method for Classification of Snowflakes Based on Images by a Multi-Angle Snowflake Camera Using Convolutional Neural Networks. Journal of Atmospheric and Oceanic Technology 36, 2267–2282, https://doi.org/10.1175/JTECH-D-19-0055.1 (2019).
Leinonen, J. & Berne, A. Unsupervised classification of snowflake images using a generative adversarial network and K-medoids classification. Atmospheric Measurement Techniques 13, 2949–2964, https://doi.org/10.5194/amt-13-2949-2020 (2020).
Key, C., Hicks, A. & Notaroš, B. M. Advanced Deep Learning-Based Supervised Classification of Multi-Angle Snowflake Camera Images. Journal of Atmospheric and Oceanic Technology https://doi.org/10.1175/JTECH-D-20-0189.1 (2021).
Schaer, M., Praz, C. & Berne, A. Identification of blowing snow particles in images from a Multi-Angle Snowflake Camera. The Cryosphere 14, 367–384, https://doi.org/10.5194/tc-14-367-2020 (2020).
Kleinkort, C., Huang, G.-J., Bringi, V. N. & Notaroš, B. M. Visual Hull Method for Realistic 3D Particle Shape Reconstruction Based on High-Resolution Photographs of Snowflakes in Free Fall from Multiple Views. Journal of Atmospheric and Oceanic Technology 34, 679–702, https://doi.org/10.1175/JTECH-D-16-0099.1 (2017).
Leinonen, J., Grazioli, J. & Berne, A. Reconstruction of the mass and geometry of snowfall particles from multi-angle snowflake camera (MASC) images. Atmospheric Measurement Techniques 14, 6851–6866, https://doi.org/10.5194/amt-14-6851-2021 (2021).
Garrett, T. J. et al. Orientations and aspect ratios of falling snow. Geophysical Research Letters 42, 4617–4622, https://doi.org/10.1002/2015GL064040 (2015).
Huang, G.-J., Kleinkort, C., Bringi, V. & Notaroš, B. M. Winter precipitation particle size distribution measurement by Multi-Angle Snowflake Camera. Atmospheric Research 198, 81–96, https://doi.org/10.1016/j.atmosres.2017.08.005 (2017).
Schirle, C. E. et al. Estimation of Snowfall Properties at a Mountainous Site in Norway Using Combined Radar and In Situ Microphysical Observations. Journal of Applied Meteorology and Climatology 58, 1337–1352, https://doi.org/10.1175/JAMC-D-18-0281.1 (2019).
Planat, N., Gehring, J., Vignon, E. & Berne, A. Identification of snowfall microphysical processes from Eulerian vertical gradients of polarimetric radar variables. Atmospheric Measurement Techniques 14, 4543–4564, https://doi.org/10.5194/amt-14-4543-2021 (2021).
Grazioli, J. & Ghiggi, G. MASCDB: a database of images, descriptors and microphysical properties of individual snowflakes in free fall. Zenodo https://doi.org/10.5281/zenodo.5578921 (2021).
Fitch, K. E., Hang, C., Talaei, A. & Garrett, T. J. Arctic observations and numerical simulations of surface wind effects on Multi-Angle Snowflake Camera measurements. Atmospheric Measurement Techniques 14, 1127–1142, https://doi.org/10.5194/amt-14-1127-2021 (2021).
Grazioli, J. et al. Measurements of precipitation in Dumont d’Urville, Adélie Land, East Antarctica. The Cryosphere 11, 1797–1811, https://doi.org/10.5194/tc-11-1797-2017 (2017).
Grazioli, J. et al. Katabatic winds diminish precipitation contribution to the Antarctic ice mass balance. Proceedings of the National Academy of Sciences 114, 10858–10863, https://doi.org/10.1073/pnas.1707633114 (2017).
Smith, C. D. et al. Evaluation of the WMO Solid Precipitation Intercomparison Experiment (SPICE) transfer functions for adjusting the wind bias in solid precipitation measurements. Hydrology and Earth System Sciences 24, 4025–4043, https://doi.org/10.5194/hess-24-4025-2020 (2020).
Genthon, C. et al. Precipitation at Dumont d’Urville, Adélie Land, East Antarctica: the APRES3 field campaigns dataset. Earth System Science Data 10, 1605–1612, https://doi.org/10.5194/essd-10-1605-2018 (2018).
Gehring, J. et al. Radar and ground-level measurements of precipitation collected by the École Polytechnique Fédérale de Lausanne during the International Collaborative Experiments for PyeongChang 2018 Olympic and Paralympic winter games. Earth System Science Data 13, 417–433, https://doi.org/10.5194/essd-13-417-2021 (2021).
Gehring, J. et al. The influence of orographic gravity waves on precipitation during an atmospheric river event at Davis, Antarctica. preprint, Atmospheric Sciences. https://doi.org/10.1002/essoar.10507051.1 (2021).
Georgakaki, P. et al. On the drivers of droplet variability in alpine mixed-phase clouds. Atmospheric Chemistry and Physics 21, 10993–11012, https://doi.org/10.5194/acp-21-10993-2021 (2021).
Gorodetskaya, I. V. et al. Cloud and precipitation properties from ground-based remote-sensing instruments in East Antarctica. The Cryosphere 9, 285–304, https://doi.org/10.5194/tc-9-285-2015 (2015).
Rew, R. & Davis, G. NetCDF: an interface for scientific data access. IEEE Computer Graphics and Applications 10, 76–82, https://doi.org/10.1109/38.56302 (1990).
Locatelli, J. D. & Hobbs, P. V. Fall speeds and masses of solid precipitation particles. Journal of Geophysical Research 79, 2185–2197, https://doi.org/10.1029/JC079i015p02185 (1974).
Heymsfield, A. J. & Kajikawa, M. An improved approach to calculating terminal velocities of plate-like crystals and graupel. Journal of Atmospheric Sciences 44, 1088–1099, https://doi.org/10.1175/1520-0469(1987)044. Place: Boston MA, USA Publisher: American Meteorological Society (1987).
Mitchell, D. L., Zhang, R. & Pitter, R. L. Mass-dimensional relationships for ice particles and the influence of riming on snowfall rates. Journal of Applied Meteorology and Climatology 29, 153–163, https://doi.org/10.1175/1520-0450(1990)029. Place: Boston MA, USA Publisher: American Meteorological Society (1990).
Mitchell, D. L. Use of mass- and area-dimensional power laws for determining precipitation particle terminal velocities. Journal of Atmospheric Sciences 53, 1710–1723, https://doi.org/10.1175/1520-0469(1996)053. Place: Boston MA, USA Publisher: American Meteorological Society (1996).
Schmitt, C. G. & Heymsfield, A. J. The Dimensional Characteristics of Ice Crystal Aggregates from Fractal Geometry. Journal of the Atmospheric Sciences 67, 1605–1616, https://doi.org/10.1175/2009JAS3187.1 (2010).
Rees, K. N., Singh, D. K., Pardyjak, E. R. & Garrett, T. J. Mass and density of individual frozen hydrometeors. Atmospheric Chemistry and Physics 21, 14235–14250, https://doi.org/10.5194/acp-21-14235-2021 (2021).
Mason, S. L., Chiu, C. J., Hogan, R. J., Moisseev, D. & Kneifel, S. Retrievals of Riming and Snow Density From Vertically Pointing Doppler Radars. Journal of Geophysical Research: Atmospheres 123, https://doi.org/10.1029/2018JD028603 (2018).
Larsen, A. B. L., Sønderby, S. K., Larochelle, H. & Winther, O. Autoencoding beyond pixels using a learned similarity metric. Available at http://arxiv.org/abs/1512.09300 (2015).
Chen, X. et al. InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets. Available at https://arxiv.org/abs/1606.03657 (2016).
Donahue, J., Krähenbühl, P. & Darrell, T. Adversarial feature learning. Available at http://arxiv.org/abs/1605.09782 (2016).
Donahue, J. & Simonyan, K. Large scale adversarial representation learning. Advances in Neural Information Processing Systems 32, 1–32 (2019).
Lipton, Z. C. & Tripathi, S. Precise recovery of latent vectors from generative adversarial networks. Available at http://arxiv.org/abs/1702.04782 (2017).
Higgins, I. et al. beta-VAE: Learning basic visual concepts with a constrained variational framework. International Conference on Learning Representations, ICLR 2017 (2017).
Dupont, E. Learning disentangled joint continuous and discrete representations. Advances in Neural Information Processing Systems 710–720 (2018).
Kim, H. & Mnih, A. Disentangling by factorising. Available at http://arxiv.org/abs/1802.05983 (2018).
Liu, X., Sanchez, P., Thermos, S., O’Neil, A. Q. & Tsaftaris, S. A. A tutorial on learning disentangled representations in the imaging domain. Available at http://arxiv.org/abs/2108.12043 (2021).
van der Maaten, L. & Hinton, G. Visualizing data using t-sne. Journal of Machine Learning Research 9, 2579–2605 (2008).
McInnes, L., Healy, J. & Melville, J. UMAP: Uniform manifold approximation and projection for dimension reduction. Available at http://arxiv.org/abs/1802.03426 (2018).
Sainburg, T., McInnes, L. & Gentner, T. Q. Parametric UMAP embeddings for representation and semisupervised learning. Available at http://arxiv.org/abs/2009.12981, https://doi.org/10.1162/neco_a_01434 (2021).
Frenzel, M. F., Teleaga, B. & Ushio, A. Latent space cartography: Generalised metric-inspired measures and measure-based transformations for generative models. Available at http://arxiv.org/abs/1902.02113 (2019).
Moon, K. R. et al. Visualizing structure and transitions in high-dimensional biological data. Nature Biotechnology 37, 1482–1492, https://doi.org/10.1038/s41587-019-0336-3 (2019).
Sainburg, T., Thielk, M. & Gentner, T. Q. Finding, visualizing, and quantifying latent structure across diverse animal vocal repertoires. PLOS Computational Biology 16, e1008228, https://doi.org/10.1371/journal.pcbi.1008228 (2020).
Aljalbout, E., Golkov, V., Siddiqui, Y., Strobel, M. & Cremers, D. Clustering with deep learning: Taxonomy and new methods. Available at http://arxiv.org/abs/1801.07648 (2018).
McConville, R., Santos-Rodriguez, R., Piechocki, R. J. & Craddock, I. N2D: (not too) deep clustering via clustering the local manifold of an autoencoded embedding. Available at http://arxiv.org/abs/1908.05968 (2019).
Mishra, D., Jayendran, A. & P., P. A. Effect of the latent structure on clustering with GANs. Available at http://arxiv.org/abs/2005.02435, https://doi.org/10.1109/LSP.2020.2996935 (2020).
Min, E. et al. A survey of clustering with deep learning: From the perspective of network architecture. IEEE Access 6, 39501–39514, https://doi.org/10.1109/ACCESS.2018.2855437 (2018).
Mukherjee, S., Asnani, H., Lin, E. & Kannan, S. ClusterGAN: Latent space clustering in generative adversarial networks. Available at http://arxiv.org/abs/1809.03627 (2018).
Shah, S. A. & Koltun, V. Deep continuous clustering. Available at http://arxiv.org/abs/1803.01449 (2018).
Wittek, P. Somoclu: An efficient distributed library for self-organizing maps. CoRR (2013).
Chen, W. et al. Deep image retrieval: A survey. arXiv 1–20 (2021).
Dubey, S. R. A decade survey of content based image retrieval using deep learning. IEEE Transactions on Circuits and Systems for Video Technology 1–1, https://doi.org/10.1109/TCSVT.2021.3080920 (2021).
Goodfellow, I. J. et al. Generative adversarial networks. Available at https://arxiv.org/abs/1406.2661. ArXiv: 1406.2661 [stat.ML] (2014).
Arjovsky, M., Chintala, S. & Bottou, L. Wasserstein GAN. Available at http://arxiv.org/abs/1701.07875 (2017).
Creswell, A. et al. Generative adversarial networks: An overview. IEEE Signal Processing Magazine 35, 53–65, https://doi.org/10.1109/MSP.2017.2765202 (2018).
Goodfellow, I. et al. Generative adversarial networks. Communications of the ACM 63, 139–144, https://doi.org/10.1145/3422622 (2020).
Saxena, D. & Cao, J. Generative adversarial networks (GANs): Challenges, solutions, and future directions. Available at http://arxiv.org/abs/2005.00065 (2020).
Deng, Y., Loy, C. C. & Tang, X. Aesthetic-driven image enhancement by adversarial learning. Available at http://arxiv.org/abs/1707.05251, https://doi.org/10.1145/3240508.3240531 (2017).
Zhao, H., Gallo, O., Frosio, I. & Kautz, J. Loss functions for image restoration with neural networks. IEEE Transactions on Computational Imaging 3, 47–57, https://doi.org/10.1109/TCI.2016.2644865 (2017).
Lehtinen, J. et al. Noise2Noise: Learning image restoration without clean data. Available at http://arxiv.org/abs/1803.04189 (2018).
Tian, C. et al. Deep learning on image denoising: An overview. Neural Networks 131, 251–275, https://doi.org/10.1016/j.neunet.2020.07.025 (2020).
Zhang, Y., Tian, Y., Kong, Y., Zhong, B. & Fu, Y. Residual dense network for image restoration. IEEE Transactions on Pattern Analysis and Machine Intelligence 43, 2480–2495, https://doi.org/10.1109/TPAMI.2020.2968521 (2021).
Wu, J., Zhang, C., Xue, T., Freeman, W. T. & Tenenbaum, J. B. Learning a probabilistic latent space of object shapes via 3D generative-adversarial modeling. Available at http://arxiv.org/abs/1610.07584 (2016).
Nguyen-Phuoc, T., Li, C., Theis, L., Richardt, C. & Yang, Y.-L. HoloGAN: Unsupervised learning of 3D representations from natural images. In 2019 IEEE/CVF international conference on computer vision workshop (ICCVW), 2037–2040, https://doi.org/10.1109/ICCVW.2019.00255 (IEEE, 2019).
Van Rossum, G. & Drake, F. L. Python 3 reference manual (CreateSpace, Scotts Valley, CA, 2009).
Pandas. https://doi.org/10.5281/zenodo.3509134 (2020).
Hoyer, S. & Hamman, J. J. xarray: N-d labeled arrays and datasets in python. Journal of Open Research Software 5, 1–6, https://doi.org/10.5334/jors.148 (2017).
Rocklin, M. Dask: Parallel computation with blocked algorithms and task scheduling. In Proceedings of the 14th python in science conference, 126–132, https://doi.org/10.25080/Majora-7b98e3ed-013 (2015).
Waskom, M. seaborn: statistical data visualization. Journal of Open Source Software 6, 3021, https://doi.org/10.21105/joss.03021 (2021).
Hunter, J. D. Matplotlib: A 2D graphics environment. Computing in Science & Engineering 9, 90–95, https://doi.org/10.1109/MCSE.2007.55 (2007).
Ramelli, F. et al. Influence of low-level blocking and turbulence on the microphysics of a mixed-phase cloud in an inner-Alpine valley. Atmospheric Chemistry and Physics 21, 5151–5172, https://doi.org/10.5194/acp-21-5151-2021 (2021).
Acknowledgements
Jacopo Grazioli has received financial support from the Swiss National Science Foundation (grant no. 200020_175700/1). This project has received support from the European Union’s Horizon 2020 research and innovation programme under grant agreement No 824310 (ICE GENESIS project). We are thankful to all the people who deployed, maintained and collected the MASC data through the years: Christophe Praz, Alfonso Ferrone, Josué Gehring, Jussi Leinonen at EPFL; Yves-Alain Roulet, Jacques Grandjean at MeteoSwiss; Claudio Duran-Alarcon at the University of Grenoble. The environmental data added in the database have been collected from multiple external or internal sources and we thank MeteoSwiss, MétéoFrance, the SLF of Davos (CH) for the data originally distributed by their operational networks. We thank prof. Tim Garrett and one anonymous reviewer for the insightful suggestions.
Author information
Authors and Affiliations
Contributions
J.G. compiled the database with input from G.G. G.G. designed the codes and packages used for the analysis (codes described and provided in this paper), with the collaboration of J.G. A.-C.B.-R. tested code and data and provided feedbacks. A.-C.B.-R. conducted the microphysical analysis shown in the manuscript. J.G. wrote the paper with input, contributions and reviews from all the authors. A.B. collected, stored and organized all the data and supervised the research.
Corresponding authors
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Grazioli, J., Ghiggi, G., Billault-Roux, AC. et al. MASCDB, a database of images, descriptors and microphysical properties of individual snowflakes in free fall. Sci Data 9, 186 (2022). https://doi.org/10.1038/s41597-022-01269-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41597-022-01269-7
This article is cited by
-
Imaging-based 3D particle tracking system for field characterization of particle dynamics in atmospheric flows
Experiments in Fluids (2023)