Article | Open | Published:

High resolution mapping of a cold water coral mound

Abstract

Cold-water coral (CWC) mounds are biogenic, long-lived morphostructures composed primarily by scleractinian CWC’s and hemipelagic sediments that form complex deep-sea microhabitats found globally but specifically along the European-Atlantic margin. In this work, high-resolution mapping was applied to identify individual organismal distribution and zonation across a CWC Piddington Mound within the Porcupine Seabight, Ireland Margin. Marine Object-Based Image Analysis (MOBIA) and different machine learning classification methods (decision tree, logistic regression, and deep neural network) were applied to a high-resolution (2 mm) reef-scale video mosaic and ROV-mounted multibeam data in order to provide new insights into the spatial organization of coral frameworks and environmental factors on CWC mounds. The results showed an accurate quantification of the amount of Coral Framework (14.5%; ~2% live and ~12.5% dead) and sponges (~3.5%) with heterogeneous distribution, restricted to a certain portion of the mound. This is the first object level quantification of live and dead coral framework facies and individual sponges across an entire CWC mound. This approach has application for habitat and conservation studies, provides a quantification tool for carbon budget assessments and can provide a baseline to assess CWC mound change. The approach can also be modified for application in other habitats.

Introduction

The extent of geographical range and ecological functioning of submarine benthic habitats are still poorly understood1,2,3. It is estimated that less than 5% of the seafloor is mapped at a resolution to that of similar studies on land4. Given the inherent difficulties of seabed mapping, specifically in relation to acquisition and analysis of marine information, new methods have now been proposed to increase knowledge from this unmapped part of the planet5,6. These include the use of structure from motion derived 3D photogrammetric reconstructions of deep-water habitats and the use of multifrequency multibeam backscatter for improved interpretations of subtle seabed features7,8.

For cold water coral (CWC) reefs and mounds, some early mapping efforts utilised regional-scale side scan sonar or multibeam echo sounder to investigate bioconstruction morphologies and seabed processes9,10,11,12,13. Later, predictive modelling and habitat suitability modelling were employed on these habitats which outlined their distribution over large areas and pointed out the need for more local-scale studies14,15,16. More recently, photogrammetry, ROV- and AUV-mounted multibeam mapping have revealed their local-scale distribution on relatively flat areas17 to near-vertical canyon walls18,19.

Despite this potential improvement of mapping seafloor habitats with coupled “video/sonar” data, seabed optical images derived from photo/video cameras mounted on Remote Operated Vehicles (ROVs) in many cases, remain limited by exploratory survey designs20,21,22, ground truthing23,24,25 and rapid ecological assessment26. In particular, efforts to map deep-water, CWCs have achieved considerable development after the use of integrated multi data spatial analysis11,27,28,29.

One of the difficulties of integrating multibeam (e.g. bathymetry and backscatter) with image data is the problem of spatial scale sensitivity to facies and habitat classification. For example, combining images of different resolution tends to increase internal variability and noise within classes and therefore may decrease the classification accuracy of traditional per-pixel basis methods30.

The concept of Geographic Object-Based Image Analysis (GEOBIA) was proposed in order to overcome problems of noise and misclassification31,32,33,34,35 and is particularly well-suited to the analysis of very high resolution (VHR) images where the increased heterogeneity of sub-meter pixels would otherwise confuse pixel-based classifications36. OBIA consists of two inter-related steps: segmentation and classification. The segmentation step is based on the creation of semantically and “meaningful” objects (polygons) based on groups of neighboring pixels with similar spectral and spatial properties. The second step is object classification which consists of allocating each object to some preselected classes based on its spectral, textural, spatial and topological characteristics. OBIA classification can offer a methodological framework for machine-learning methods which takes into account multiple properties of image objects31,37.

Diesing38 adapted the term to Marine Object-Based Image Analysis (MOBIA) as the application of GEOBIA to marine data sets with the aim of mapping seafloor geomorphology, geology and habitats. The use of MOBIA, however, is still lacking in more widespread applications. There are some explanations for this, firstly, the spatial distribution of the dynamics of marine and ocean phenomena (entities and processes) do not conform to patterns established with those from terrestrial areas. Furthermore, there are no robust, established theoretical models that can be used to characterize the spatial hierarchy of marine phenomena, especially in relation to ecological interactions, which could guide the establishment of spatial units based, for example, on watershed analysis or Landscape Ecology (see39). Manderson40, in addressing this question, indicates that the basic analysis of marine environments following an ecosystem approach should always be associated with the scale of the dominant hydrodynamic process. The derived “Benthoscapes” in this context would reflect a set of spatially enclosed habitats controlled by a specific hydrodynamic process. The controlling process however could be difficult to detect by remote sensing techniques making this task difficult or subjective41.

Secondly, the original “object” modeling methodologies developed for terrestrial areas were primarily for optical multispectral satellite image analysis, which allows a direct and clear delimitation of spatial aggregation units. In the case of submarine environments, acoustic remote sensing (through multibeam sonars or side scan sonars, for example) does not add as much information; the variability in the acoustic signal is directly related to the type of substrate (texture and composition) which does not necessarily reflect direct ecological characteristics. As such, it can be said that much of the factors that characterize the variation within marine environments (or habitats) are “invisible” to acoustic signals.

Recent studies indicate that promising results for the characterization of “Benthoscapes” from the segmentation and application of OBIA of marine substrates at the meter-scale resolution can be achieved using optical data such as video-imaging (examples in42,43). In the case of optical-scale surveys, in which the variation of environments is evident such as with coral reefs, object-oriented classification has shown even more consistent results44,45. A pioneer study by Purser et al.46, showed that machine learning of optical-scale survey data provided accurate estimation of live cold water coral densities.

The objective of this study is to; i) develop and compare a supervised classification of a high-resolution, reef-scale, deep-water coral image mosaic of the Piddington Mound (Porcupine Seabight, NE Atlantic) using MOBIA and a machine learning classification (MLC) to an established manual classification and; ii) establish the most appropriate machine learning classification for integrated ROV bathymetric and video datasets. In a broader context, we use the findings of this work to provide unique insights into the spatial organization of coral frameworks and environmental factors on CWC mounds. The results of this work can be used as a quantified baseline to which other coral mound surfaces can be compared. For instance, this whole reef approach would enable more accurate quantifications for carbon budget analyses and would allow time series study to assess changes in the reefs. Furthermore, the MLC applied here has a broad significance as it can easily be shared, customised and applied to other deep- and shallow-water habitats.

Materials and Methods

Study Site

The Piddington Mound, a CWC mound in the Belgica Mound Province (BMP), has been selected for this study due to the existence of high-resolution imagery (video and bathymetry) which covers the entire surface of the mound as presented in29. The BMP is a Special Area of Conservation (SAC) located within the Porcupine Seabight, NE Atlantic (Fig. 1) that is known for the abundance of “giant” CWC mounds, up to 100 m in height, broadly aligned as two contour-parallel mound chains13. Smaller CWC mounds approximately 10 m tall, called the Moira Mounds, exist between these giant mounds47,48. The Piddington Mound is one of these smaller CWC mounds with a spatial extent of approximately 40 × 60 m and a current-aligned, ovoid morphology. Speculated to be Holocene in age49, the Moira Mounds are predominantly distributed across 4 areas; the northern area, the upslope area, the downslope area and the midslope area. The Piddington Mound exists in the downslope area, described as favorable for mound development with current speeds of approximately 40 cm s−1. Glacially-derived dropstones and fine hemipelagic sediments exist in the off-mound area, surrounding the Piddington Mound, while the mound itself is covered predominantly by Lophelia pertusa, Madrepora oculata colonies, other marine organisms such as sponges and echinoderms as well as coral rubble and sediments29.

Figure 1
figure1

Study site area. The Piddington Mound bathymetry in the context of the Moira Mounds region derived from ROV-based multibeam echosounder data.

Data

ROV-video data were collected for this research during the VENTuRE survey (2011) on board RV Celtic Explorer with the Holland 1 ROV (cruise number CE11009). The video data were acquired across the Piddington Mound using a downward-facing HD camera which was mounted at the bottom of the ROV. Positioning and navigation data for the ROV during the dive were recorded using a Sonardyne Ranger 2 USBL (ultra-short baseline positioning system) corrected by an RDI Workhouse doppler velocity logger. The ROV altimeter recorded the height of the ROV (and therefore camera) above the seabed. Downward-facing HD video data were recorded during a series of transects covering the entire surface of the Piddington Mound. To maintain a clear image of the mound surface, the ROV was kept <2 m above the mound surface at a survey speed of approximately 0.8 knots. Several lights were attached to the ROV to achieve homogenous lighting across the camera field of view.

The ROV-video dataset was converted into a video mosaic (Lim et al.29) using the IFREMER in-house software Matisse which extracted images from the raw video data at a rate of 1 per second. Poor quality video data (imagery collected more than 2 m above the seabed, collected in poor water quality, acquired too quickly or with poor navigation) were excluded from the image extraction process. To lower the trajectory noise of the ROV, sliding median filtering and 2nd order polynomial model fit was applied to all the USBL navigation data. The extracted images and filtered navigation were synchronized so that each image had an approximate position, later refined by the mosaicking process. Features in the extracted images were detected and matched using a SIFT (Scale Invariant Feature Transform) algorithm17. The resulting matched images and USBL navigation were merged to give an accurate global position, correct scaling and sufficient local overlapping through a cost function minimization. The final mosaic was projected within a GIS (In UTM 29N) and resampled to a 2 mm resolution. The boundary of the mound was defined by a topographical break (derived from the DTM), above which the inclination of the slope increases significantly (Fig. 2).

Figure 2
figure2

Video mosaic of the Piddington Mound. The mound area delineated by topographic break is marked by the yellow line.

A Kongsberg EM2040 multibeam echosounder was mounted to the Holland 1 ROV where it was integrated with a sound velocity sensor, Kongsberg HAIN inertial navigation system. The ROV was flown at approximately 30 m above the seabed around the Piddington Mound where multibeam data acquisition was managed and monitored via Seafloor Information System (SIS). The multibeam was operated at 400 kHz at a survey speed of 0.8 knots until data were acquired from the Piddington Mound and surrounding seabed. All data were saved as *.all files. The files were imported to Qimera where a zero tide was applied and all lines were gridded together. Using the swath editor, each line was manually inspected for anomalous data spikes, which were manually removed. The cleaned multibeam data were gridded at a resolution of 0.5 m and saved in an ArcGRID format.

Video Mosaic Segmentation and Classification

The video mosaic segmentation process was performed to the whole video mosaic area using the multi-resolution segmentation algorithm in software eCognition v9.050. Starting from an individual pixel, it consecutively merges pixels from the original until a certain threshold, defined by the scale parameter, is reached creating a polygon, or an “object” with similarities in scale/shape and brightness/colour51. In this way, the segmentation model parameters tuning defined by shape/colour and the scale parameter definition can be quite subjective and dependent on trial and error and the analyser’s subjectivity52,53,54. Since the main goal of this study is to perform a zonation analysis of the Piddington Mound and its main components, the optimum segmentations should be detailed enough to define units (individuals) of key biotic (e.g. coral frameworks, sponges and echinoderms) and seabed features (such as pebbles or sediment patches).

Figure 3 shows examples of different segmentation strategies on the same areas (a central sector of the study region). Figure 3a illustrates an over-segmented section where too many objects were created. In contrast, Fig. 3c shows an under-segmented area where few objects are delineated thereby failing to reflect the bio-geophysical characteristics of the whole diversity of seabed features55. In order to find an optimum parameterization of the segmentation model, an Estimation Scale Parameter model (ESP), developed by56 was used which defines the empirical relationship between spatial structures of the image and the size of created objects. Based on the calculated results of the ESP model for the area, the Scale Parameter was defined as “400” producing an apparently adequate segmentation (i.e. with the features such as coral frameworks and sponges’ polygons were clearly delimited - Fig. 3b).

Figure 3
figure3

Segmentation differences for part of the study area showing examples of over segmentation. (a) Scale Parameter 50 (over segmentation). (b) Scale Parameter 400 (defined by the ESP Tool). (c) Scale Parameter 1000 (under segmentation).

For each polygon (object) defined by the segmentation tool, a set of features were calculated relating to shape, form, texture and context. At this phase, bathymetric data were also incorporated to the dataset. Table 1 presents the details of the selected features. Detailed calculations of these texture measures can be found in52.

Table 1 Spatial features used as parameters in classification process.

The next step was the classification process which assigns map categories to each segment using membership rules57. It defined hierarchical class labels based on three distinct categories: (1) Biogenic; (2) Sediment and; (3) other/non classified. The Biogenic category was subdivided in “live coral framework”, “dead coral framework” and “sponges”. The Sediment category was subdivided in “hemipelagic sediment” (sand or mud with no recognisable bioclasts or dropstones), “hemipelagic sediment with dropstones” and “dropstones/pebbles”. The category “other/non classified” was related to objects with few representatives in the overall environment (e.g. fish, crab, echinoderms) or with no significance (e.g. shadows). The categories were based on the work of29 in the same area (the Piddington Mound). Figure 4 shows examples of selected categorical objects of each class label.

Figure 4
figure4

Segments (objects) based on self-existent and resoluble entities: biogenic divided into (a) living coral framework (mainly Lophelia pertusa and Madrepora oculata); (b) glass sponges; (c) coral rubble and; (d) dead coral framework. Sediment divided into (e) hemipelagic sediments; (f) hemipelagic sediment with dropstones and; (g) pebbles. Non-classified divided into (h) shadows; (i) fish and; (j) echinoderm/others. The yellow bar represents a 20 cm scale reference.

150 training samples were selected for each significance category except the “other” category which was not considered for this analysis due to its rare occurrence and insignificance in the area as a whole. Three different supervised classification methods were applied: Log Regression (LR), Random Forest (RF) and Deep Neural Network (DNN). The training and classification models were developed in Microsoft Azure Machine Learning Studio (MAMLS)58 and implemented in Python 3.4.

LR Multiclass supervised classification is an extension of binary logistic regression which categorizes objects based on their closest training samples in feature space predicting class probability based on the input features after ranking them according to their relative importance59,60,61. Random Forest is a method that operates by constructing multiple decision trees (i.e. classification trees, where the leaves represent classifications and the branches represent conjunctions of features that produce those classifications). Then by voting for the highest output class, it searches away from the unknown object to be classified in all directions until it encounters k user-specified training objects and assigns the object to the class with the majority vote of the encountered objects62,63. The DNN method is inspired by the way biological nervous systems process information. It consists of a set of interconnected layers, in which the inputs lead to outputs by a series of weighted elements (edges and nodes of Neural Networks). A particular set of neural network algorithms, made up of more than three layers along with the input, output and more than one hidden layer, are known as “Deep” learning algorithms with the method itself is referred to as a “Deep Neural Network”64,65.

As the living biogenic facies (i.e. living corals and sponge objects) covers quite a small individual area in relation to the total area, it can be difficult to characterize the distribution of these classes in the context of the mound. In order to better describe the spatial distribution of these classes, “hot spot” analysis using Getis-Ord GI* function was performed. This tool works by looking at each feature within the context of neighboring features (see66,67). For all methods, classification model frameworks were built based on data preparation, training, model creation using DNN, RT and LR algorithms and model accuracy evaluation. The output classification results were exported into a geodatabase for subsequent spatial analysis and map preparation (see Fig. 5).

Figure 5
figure5

Methodology workflow proposed for the Piddington Mound classification.

Results

Accuracy assessment

For the analysis of class prediction, a final accuracy assessment of 15% of the dataset were sampled randomly for training and the remaining 85% were used as test set. The multi-class classification performance matrix taken from MAMLS model set is shown in Fig. 6. The precision and recall matrices were assessed by two different metrics: micro-averaging and macro-averaging. Micro-averaging tends to be effective in the most frequent classes whereas macro-averaging considers each class equally. The experimental result of the models are presented in the Table 2.

Figure 6
figure6

Accuracy matrix for classification methods LR, RF and DNN: Classes:LCN) Live Coral Framework; DC) Dead Coral Framework; SPG) Sponges; HEMS Hemipelagic Sediments; HESDR) Hemipelagic Sediments with Dropstones; CRUB) Coral Rubble; and PEB) Pebbles.

Table 2 Accuracy metrics for the classification methods.

Classification

The classified maps generated by the three classification models are presented in Figs 7, 8 and 9 and the area and object counting summary is presented in Table 3. The general trend of the class distribution is quite similar in the three methods, with overall accuracies between 87% and 83% (Table 2). Although the RF and LR classifiers yielded the lowest general results, they allow the interpretation and control of their parameters with adjustments and rules settings. In contrast, DNN are more complex where interpretation is more difficult and can only be verified externally37.

Figure 7
figure7

Classified Map - Log Regression (LCN) Live Coral Framework; (DC) Dead Coral Framework; (SPG) Sponges; (HEMS) Hemipelagic Sediments; (HESDR) Hemipelagic Sediments with Dropstones; (CRUB) Coral Rubble; (PEB) Pebbles and (SHAD) Other/Shadows.

Figure 8
figure8

Classified Map - Random Forrest (LCN) Live Coral Framework; (DC) Dead Coral Framework; (SPG) Sponges; (HEMS) Hemipelagic Sediments; (HESDR) Hemipelagic Sediments with Dropstones; (CRUB) Coral Rubble; (PEB) Pebbles and (SHAD) Other/Shadows.

Figure 9
figure9

Classified Map – Deep Neural Network (LCN) Live Coral Framework; (DC) Dead Coral Framework; (SPG) Sponges; (HEMS) Hemipelagic Sediments; (HESDR) Hemipelagic Sediments with Dropstones; (CRUB) Coral Rubble; (PEB) Pebbles and (SHAD) Other/Shadows.

Table 3 Areas and object statistics – % T.A. (% of object in total area), # of Obj. (total number of objects) M.O.A. (mean size of each object, in m2) % M.A. (% of area inside the mound limits).

For the “sediment cover” classes (i.e. “Coral Rubble” and “Hemipelagic sediments”), the accuracy analysis (Fig. 6) shows a significative drop in statistical confidence especially in LR and RF methods for both classes. The DNN method however, maintained an accuracy higher than 80% for these classes indicating a better discrimination capability between such classes. Visual inspection confirmed that the results of the DNN were more effective, although in some cases it can be difficult to differentiate between classes even visually.

For smaller and more distinct objects such as sponges and living corals, the results were less uniform with an accuracy of >80% (except in RF classification for sponges which had a confidence of approx. 70% where the classification has an issue discerning between living and dead coral framework).

The zonation trend of the mound is quite clear with a prevalence of hemipelagic sediment with dropstones and larger pebbles on the deeper, southern part of the mound while at the northern, marginal areas the classes “hemipelagic sediment” and “coral rubble” are more dominant. The mound itself is covered by biogenic facies “dead coral framework” and “coral rubble”. At the highest parts of the mound (below −974 m depth), both classes represent more than 60% of the coverage (against approx. 36% in overall area).

Living biogenic facies (i.e. living corals and sponge objects) have a relatively small individual area and can be difficult to identify in the general map context. As such, the “hot spot” map (Fig. 10) shows that high concentrations of living corals are located to the north face of the mound, with very few occurrences outside the 99% confidence area. Glass sponges clearly occur and are related to both the mound summit and the upper western face of the mound (Fig. 11) (note, the sponges class is represented by dots due to the small dimensions of polygons). The scatter-plot graphs (Fig. 12) show a clear correlation between Z-score of hotspot cells and the bathymetry of sponges (a) indicating a high correlation of sponge occurrence and topographical settings. while the trend is much less evident among corals (12 b).

Figure 10
figure10

Hot Spot map of coral object distribution.

Figure 11
figure11

Hot Spot Map of sponges’ object distribution.

Figure 12
figure12

Scatter plot for between Z-score of glass sponges (a) and coral (b) hotspot cells (with colour scale related to Figs 10 and 11) and the bathymetry.

The polar plot of the Living coral and Sponges (Fig. 13) shows the distribution of living corals and sponges in relation to slope orientation or aspect. It is very clear that the arrangement of living organisms (sponges and corals) on the mound obeys a distribution pattern restricted to the northern mound flank. In the case of corals, there are two oppositely symmetrical directions: approximately 300° (WNW) and 70° (ENE). A similar but less tight relationship between occurrence and aspect is also revealed for sponges.

Figure 13
figure13

Polar plot indicating mound slope orientation where coral and sponges have settled.

Discussion

The quality and accuracy of object-based image analysis applied to CWC mound photo mosaicking as described in this paper are dependent on two steps: Segmentation and Machine learning classification algorithms. The segmentation process is dependent on empirical calibration, the basis of a trial and error procedure where the optimum level of feature discrimination can be reached. New methods of Object-based scale parameter selection have been proposed and applied38,68,69 but Kim et al.70 noted that defining the most suitable scale for image segmentation is still quite problematic. In this case, we applied a local variance (LV) of object heterogeneity to determine the best scale parametrization as proposed in56. However, in some cases, areas with sponges, pebbles and smaller objects such as isolated corals and echinoderms were under-segmented when they existed over complex backgrounds (such as coral rubble or dead coral framework). The diminution of scale parameter, yielding the creation of detailed segments, resulted in a considerable increase in processing time and over-segmentation of objects which decreased the accuracy of the classification algorithms.

The Machine learning classification models have produced satisfactory results by creating facies distribution maps of the Piddington mound with improved results compared to the manual counting method described by Lim et al.29. Some advantages of the ML object classification can be highlighted: (1) processing and modelling time is considerably lower especially after tuning the segmentation and classification algorithm parameters; (2) it offers a less-subjective and observer-dependent approach to mapping and classifying coral facies; (3) it can be specific to typical coral facies quantification and characterization (i.e. sponges, coral framework) by tuning the parameters to better perform identification (segmentation) and classification of these features. Subsequently, the algorithm can be trained to compare different sites and therefore indicate quantitative structural differences between frameworks (i.e. mound, reefs); (4) it produces an ordinary georeferenced file in a geodatabase format (e.g. “shapefile”) that can aggregate more information into object fields such as taxonomic or sampling data which can increase the accuracy of new classification processes and; (5) it allows the manipulation of classes (signing, recombining, merging).

In terms of overall classification, all three methods tested herein produced similar results to map mound facies. For living organism classes (living coral and sponges) the DNN method showed a better discrimination performance in terms of accuracy level as well as with a lower misclassification error (mainly related to differentiation between living and dead coral framework). There was a relevant classification confusion between sponges and living corals in all methods. Visual inspection of these misclassified objects indicates that since sponges almost always appear in isolated occurrences (and the samples were chosen in this condition), when they occurred in pairs or more adjacent individuals, the models invariably classify them as “living corals”. This issue can be addressed by “re-segmentation” of the area with a lower scale parameter. This would help to identify individual organisms as smaller segments (objects) however, it is important to note that this may lead to a lower classification performance in other classes.

In the case of non-living classes, the major problems of classification occurred in the differentiation of allogenic and autogenic carbonate-dominated substrates (coral rubble and dead coral framework) (see Fig. 6). The DNN method was able to differentiate between these classes which can be evidenced at the top of the mound where there are significantly more “rubble coral” classes in the DNN model map reflected by a higher accuracy for this class (88% against 59.3% and 29.6% for RF and LR, respectively). For “sediment substrates” (Hemipelagic Sediments and Hemipelagic Sediments with dropstones) the methods also showed similar results with good discrimination rates and accuracy (>80%).

The MOBIA carried out here offers an accurate quantification of the amount of Coral Framework (12%; ~3.5% live and ~8.5% dead) and sponges (2%) across the Piddington Mound. This is the first object-level estimation of live and dead coral framework facies and individual sponges across an entire CWC mound. Interestingly ~29% of the mound surface is covered by coral rubble.

While coral rubble and dead coral framework are found across the mound, the live coral framework is restricted to the northern sector of the mound. The distribution of sponges also shows a hot spots distribution towards the highest parts of the mound, such pattern was not observed with the living corals (Fig. 12a,b). Lim et al.29, suggest that this restriction can be attributed to the high current speeds in the area and the relation to optimal conditions for coral feeding and hydrodynamics and for larvae to find hard substrates to attach to71.

Given the northern restriction of the live coral frameworks, a continued development of the mound in contemporary conditions is likely to generate an asymmetric north-south mound profile. Conversely, the mound has a north-south symmetric elongation (Fig. 1). As such, it is unlikely that the distribution of the live coral has been restricted to the northern portion of the mound throughout its development. Furthermore, given the occurrence of dead coral framework across the full mound which has not yet been buried by sediment or bioeroded due to long exposure, this suggests that the northern restriction is related to a recent change in environmental conditions. This suggestion can also be strengthened by recent observations which show that there has been a total of 19% change in the proportion of sediment (hemipelagic and bioclastic) and coral frameworks (live and dead) on the Piddington Mound surface from 2011 to 201522.

Conclusion

Three different machine learning classification methods (decision tree, logistic regression, and multilayer deep neural network) were applied to a high-resolution, segmented, reef-scale video mosaic and ROV-mounted multibeam data. The results show that Object-Based Image Analysis (OBIA) derived from the grouping of similar pixels in “objects” (self-existent and resoluble entities) with similar characteristics was particularly successful when applied to high resolution marine habitat mapping, specifically to cold water coral mound facies. Further, the concept of “MOBIA” (Marine OBIA) might be used since it considers not only optical imaging, but also acoustic data to segment and classify seabed features. Such techniques proved considerably more effective than manual and/or pixel-based approaches.

In order to perform MOBIA, segmentation and classification methods should be applied. Although some models for estimating segmentation parameters exist, the process of defining optimal characteristics of objects (according to seabed variables) is still largely empirical and analyst-dependent. New Machine Learning classification methods, widely available from private and open source platforms, has proven to be quite successful in terms of mapping accuracy. Deep Neural Networks showed an overall higher classification accuracy, although the Random Forest and Log Regression showed similar results.

Given the spatial coverage (100% of the CWC mound) and data resolution (2 mm video mosaic and 10 cm bathymetry), MOBIA was applied to an entire CWC mound for the first time to quantify individual organisms (e.g. sponges), coral framework coverage and typical sedimentary facies. The results show that the mound has a high coverage of coral rubble (29%) and only 12% of the mound was covered by coral framework.

References

  1. 1.

    Snelgrove, P. V. The biodiversity of macrofaunal organisms in marine sediments. Biodiversity & Conservation 7(9), 1123–1132 (1998).

  2. 2.

    Williams, S. B., Pizarro, O., Jakuba, M. & Barrett, N. AUV benthic habitat mapping in south eastern Tasmania. In Field and Service Robotics. 275–284. Springer, Berlin, Heidelberg (2010).

  3. 3.

    Teixeira, J. B. et al. Traditional ecological knowledge and the mapping of benthic marine habitats. Journal of environmental management 115, 241–250 (2013).

  4. 4.

    Wright, D. J. & Heyman, W. D. Introduction to the special issue: marine and coastal GIS for geomorphology, habitat mapping, and marine reserves. Marine Geodesy 31(4), 223–230 (2008).

  5. 5.

    Wright, D. & Barlett, D. (Ed.). Marine and coastal geographical information systems. CRC press (2000).

  6. 6.

    Mayer, L. et al. The Nippon Foundation—GEBCO seabed 2030 project: The quest to see the world’s oceans completely mapped by 2030. Geosciences 8(2), 63 (2018).

  7. 7.

    Robert, K., Jones, D. O. B., Roberts, J. M. & Huvenne, V. A. I. Improving predictive mapping of deep-water habitats: Considering multiple model outputs and ensemble techniques. Deep Sea Research Part I: Oceanographic Research Papers 113, 80–89 (2016).

  8. 8.

    Feldens, P., Schulze, I., Papenmeier, S., Schönke, M. & Schneider von Deimling, J. Improved Interpretation of Marine Sedimentary Environments Using Multi-Frequency Multibeam Backscatter Data. Geosciences 8, 214 (2018).

  9. 9.

    Beyer, A., Schenke, H. W., Klenke, M. & Niederjasper, F. High resolution bathymetry of the eastern slope of the Porcupine Seabight. Marine Geology 198, 27–54 (2003).

  10. 10.

    Taviani, M. et al. First geo-marine survey of living cold-water Lophelia reefs in the Ionian Sea (Mediterranean basin). Facies, 50(3-4), 409–417 (2005).

  11. 11.

    Roberts, J. M., Brown, C. J., Long, D. & Bates, C. R. Acoustic mapping using a multibeam echosounder reveals cold-water coral reefs and surrounding habitats. Coral Reefs 24(4), 654–669 (2005).

  12. 12.

    Wheeler, A. J. et al. Morphology and environment of cold-water coral carbonate mounds on the NW European margin. International Journal of Earth Sciences 96, 37–56 (2007).

  13. 13.

    Wheeler, A. J. et al. Sedimentary processes and carbonate mounds in the Belgica Mound province, Porcupine Seabight, NE Atlantic, in: Freiwald, A., Roberts, J.M. (Eds), Cold-water Corals and Ecosystems. Springer-Verlag, Berlin Heidelberg, pp. 533–564 (2005).

  14. 14.

    Davies, A. J. & Guinotte, J. M. Global Habitat Suitability for Framework-Forming Cold-Water Corals. PLoS ONE 6, e18483 (2011).

  15. 15.

    Davies, A. J., Wisshak, M., Orr, J. C. & Roberts, J. M. Predicting suitable habitat for the cold-water coral Lophelia pertusa (Scleractinia). Deep-Sea. Research 55, 1048–1062 (2008).

  16. 16.

    Dolan, M. F., Grehan, A. J., Guinan, J. C. & Brown, C. Modelling the local distribution of cold-water corals in relation to bathymetric variables: Adding spatial context to deep-sea video data. Deep Sea Research Part I: Oceanographic Research Papers 55(11), 1564–1579 (2008).

  17. 17.

    Lim, A., Kane, A., Arnaubec, A. & Wheeler, A. J. Seabed image acquisition and survey design for cold water coral mound characterisation. Marine Geology 395, 22–32 (2018).

  18. 18.

    Huvenne, V. A. et al. A picture on the wall: innovative mapping reveals cold-water coral refuge in submarine canyon. PLoS ONE 6, e28755 (2011).

  19. 19.

    Roberts, J. M., Wheeler, A. J. & Freiwald, A. Reefs of the Deep: The Biology and Geology of Cold-Water Coral Ecosystems. Science 312, 543–547 (2006).

  20. 20.

    Parry, D. M., Kendall, M. A., Pilgrim, D. A. & Jones, M. B. Identification of patch structure within marine benthic landscapes using a remotely operated vehicle. Journal of Experimental Marine Biology and Ecology 285, 497–511 (2003).

  21. 21.

    Boavida, J., Assis, J., Reed, J., Serrão, E. A. & Gonçalves, J. M. Comparison of small remotely operated vehicles and diver-operated video of circalittoral benthos. Hydrobiologia 766(1), 247–260 (2016).

  22. 22.

    Lim, A., Huvenne, V. A. I., Vertino, A., Spezzaferri, S. & Wheeler, A. J. New insights on coral mound development from groundtruthed high-resolution ROV-mounted multibeam imaging. Marine Geology 403, 225–237 (2018).

  23. 23.

    Vertino, A. et al. Benthic habitat characterization and distribution from two representative sites of the deep-water SML Coral Province (Mediterranean). Deep Sea Research Part II: Topical Studies in Oceanography 57(5-6), 380–396 (2010).

  24. 24.

    Lacharité, M., Brown, C. J. & Gazzola, V. Multisource multibeam backscatter data: developing a strategy for the production of benthic habitat maps using semi-automated seafloor classification methods. Marine Geophysical Research 39(1-2), 307–322 (2018).

  25. 25.

    Ismaili, K., Huvenne, V. & Robert, K. Quantifying spatial heterogeneity in submarine canyons. Progress in Oceanography. 2018 – In Press, https://doi.org/10.1016/j.pocean.2018.03.006.

  26. 26.

    Tilot, V. C., Ormond, R., Navas, J. M. & Catalá, T. S. The benthic megafaunal assemblages of the CCZ (Eastern Pacific) and an approach to their management in the face of threatened anthropogenic impacts. Frontiers in Marine Science 5, 7 (2018).

  27. 27.

    Guinan, J., Grehan, A. J., Dolan, M. F. & Brown, C. Quantifying relationships between video observations of cold-water coral cover and seafloor features in Rockall Trough, west of Ireland. Marine Ecology Progress Series 375, 125–138 (2009).

  28. 28.

    Dolan, M. F. J., Grehan, A. J., Guinan, J. C. & Brown, C. Modelling the local distribution of cold-water corals in relation to bathymetric variables: Adding spatial context to deep-sea video data. Deep-Sea. Research I 55, 1564–1579 (2008).

  29. 29.

    Lim, A., Wheeler, A. J. & Arnaubec, A. High-resolution facies zonation within a cold-water coral mound: The case of the Piddington Mound, Porcupine Seabight, NE Atlantic. Marine Geology 390, 120–130 (2017).

  30. 30.

    Schiewe, J., Tufte, L. & Ehlers, M. Potential and problems of multi-scale segmentation methods in remote sensing. GeoBIT/GIS 6(01), 34–39 (2001).

  31. 31.

    Blaschke, T. Object based image analysis for remote sensing. ISPRS journal of photogrammetry and remote sensing 65(1), 2–16 (2010).

  32. 32.

    Benz, U. C., Hofmann, P., Willhauck, G., Lingenfelder, I. & Heynen, M. Multi-resolution, object-oriented fuzzy analysis of remote sensing data for GIS-ready information. ISPRS Journal of photogrammetry and remote sensing 58(3-4), 239–258 (2004).

  33. 33.

    Yan, G., Mas, J. F., Maathuis, B. H. P., Xiangmin, Z. & Van Dijk, P. M. Comparison of pixel‐based and object‐oriented image classification approaches—a case study in a coal fire area, Wuda, Inner Mongolia, China. International Journal of Remote Sensing 27(18), 4039–4055 (2006).

  34. 34.

    Bock, M., Xofis, P., Mitchley, J., Rossner, G. & Wissen, M. Object-oriented methods for habitat mapping at multiple scales–Case studies from Northern Germany and Wye Downs, UK. Journal for Nature Conservation 13(2-3), 75–89 (2005).

  35. 35.

    Hay, G. J. & Castilla, G. Geographic Object-Based Image Analysis (GEOBIA): A new name for a new discipline. In Object-based image analysis. 75–89. Springer, Berlin, Heidelberg (2008).

  36. 36.

    Leon, J. & Woodroffe, C. D. Improving the synoptic mapping of coral reef geomorphology using object-based image analysis. International Journal of Geographical Information Science 25(6), 949–969 (2011).

  37. 37.

    Peña, J. M. et al. Object-based image classification of summer crops with machine learning methods. Remote Sensing 6(6), 5019–5041 (2014).

  38. 38.

    Diesing, M. Application of geobia to map the seafloor.Proceedings GEOBIA 2016: Solutions and Synergies, University of Twente Faculty of Geo-Information and Earth Observation (ITC) (2016).

  39. 39.

    Lechner, A. M. & Rhodes, J. R. Recent progress on spatial and thematic resolution in landscape ecology. Current Landscape Ecology Reports 1(2), 98–105 (2016).

  40. 40.

    Manderson, J. P. Seascapes are not landscapes: an analysis performed using Bernhard Riemann’s rules. ICES Journal of Marine Science 73(7), 1831–1838 (2016).

  41. 41.

    Zajac, R. N. Challenges in marine, soft-sediment benthoscape ecology. Landscape Ecology 23(1), 7–18 (2008).

  42. 42.

    Siakavara, K. & Argialas, D. Automated identification of geomorphological and topographical and features of the seabed by object based image analysis of digital terrain models. Zeitschrift für Geomorphologie, Supplementary. Issues 57(3), 63–74 (2013).

  43. 43.

    Lucieer, V., Hill, N. A., Barrett, N. S. & Nichol, S. Do marine substrates ‘look’ and ‘sound’ the same? Supervised classification of multibeam acoustic data using autonomous underwater vehicle images. Estuarine, Coastal and Shelf Science 117, 94–106 (2013).

  44. 44.

    Marsh, L., Copley, J. T., Huvenne, V. A. & Tyler, P. A. Getting the bigger picture: using precision Remotely Operated Vehicle (ROV) videography to acquire high-definition mosaic images of newly discovered hydrothermal vents in the Southern Ocean. Deep Sea Research Part II: Topical Studies in Oceanography 92, 124–135 (2013).

  45. 45.

    Diesing, M. et al. Mapping seabed sediments: comparison of manual, geostatistical, object-based image analysis and machine learning approaches. Continental Shelf Research 84, 107–119 (2014).

  46. 46.

    Purser, A., Bergmann, M., Lundälv, T., Ontrup, J. & Nattkemper, T. W. Use of machine-learning algorithms for the automated detection of cold-water coral habitats: a pilot study. Marine Ecology Progress Series 397, 241–251 (2009).

  47. 47.

    Foubert, A. et al. The Moira Mounds, small cold-water coral mounds in the Porcupine Seabight, NE Atlantic: Part B—Evaluating the impact of sediment dynamics through high-resolution ROV-borne bathymetric mapping. Marine Geology 282(1–2), 65–78 (2011).

  48. 48.

    Wheeler, A. J. et al. The Moira Mounds, small cold-water coral banks in the Porcupine Seabight, NE Atlantic: Part A—an early stage growth phase for future coral carbonate mounds? Marine Geology 282, 53–64 (2011).

  49. 49.

    Huvenne, V. A. I. et al. The seabed appearance of different coral bank provinces in the Porcupine Seabight, NE Atlantic: results from sidescan sonar and ROV seabed mapping, in: Freiwald, A., Roberts, J. M. (Eds), Cold-water Corals and Ecosystems. Springer-Verlag, Berlin Heidelberg, pp. 535–569 (2005).

  50. 50.

    Trimble. Ecognition Developer 8.64. 1 User Guide (2011).

  51. 51.

    Lowe, D. G. Object recognition from local scale-invariant features. In Computer vision, 1999. The proceedings of the seventh IEEE international conference 2, 1150–1157. IEEE. 1150–1157 (1999).

  52. 52.

    Ierodiaconou, D. et al. A. Combining pixel and object-based image analysis of ultra-high resolution multibeam bathymetry and backscatter for habitat mapping in shallow marine waters. Marine Geophysical Research 39(1–2), 271–288 (2018).

  53. 53.

    Walker, J. & Blaschke, T. Object-based landcover classification for the Phoenix metropolitan area: optimization vs. transportability. International Journal of Remote Sensing 29, 2021–2040 (2008).

  54. 54.

    Gerçek, D., Toprak, V. & Strobl, J. Object-based classification of landforms based on their local geometry and geomorphometric context. International Journal of Geographical Information Science 25(6), 1011–1023 (2011).

  55. 55.

    Brown, C. J., Sameoto, J. A. & Smith, S. J. Multiple methods, maps, and management applications: purpose made seafloor maps in support of ocean management. Journal of Sea Research 72, 1–13 (2012).

  56. 56.

    Drǎguţ, L., Tiede, D. & Levick, S. R. ESP: a tool to estimate scale parameter for multiresolution image segmentation of remotely sensed data. International Journal of Geographical Information Science 24(6), 859–871 (2010).

  57. 57.

    Phinn, S. R., Roelfsema, C. M. & Mumby, P. J. Multi-scale, object-based image analysis for mapping geomorphic and ecological zones on coral reefs. International Journal of Remote Sensing, 33(12), 3768–3797 (2012).

  58. 58.

    Mund, S. Microsoft azure machine learning. Packt Publishing Ltd (2015).

  59. 59.

    Qasem, M., Thulasiram, R. & Thulasiram, P. Twitter sentiment classification using machine learning techniques for stock markets. In Advances in Computing, Communications and Informatics (ICACCI), International Conference on (pp. 834–840) IEEE (2015).

  60. 60.

    Bihis, M. & Roychowdhury, S. A generalized flow for multi-class and binary classification tasks: An Azure ML approach. In Big Data (Big Data), 2015 IEEE International Conference on (pp. 1728–1737). IEEE (2015).

  61. 61.

    Quinlan, J. R. C4. 5: programs for machine learning. Elsevier (2014).

  62. 62.

    Pal, M. Random forest classifier for remote sensing classification. International Journal of Remote Sensing 26(1), 217–222 (2005).

  63. 63.

    Zhang, C. Applying data fusion techniques for benthic habitat mapping and monitoring in a coral reef ecosystem. ISPRS Journal of Photogrammetry and Remote Sensing 104, 213–223 (2015).

  64. 64.

    Krizhevsky, A., Sutskever, I. & Hinton, G. E. Imagenet classification with deep convolutional neural networks. In Advances in neural information processing systems. 1097–1105 (2012).

  65. 65.

    Bosse, S., Maniry, D., Müller, K. R., Wiegand, T. & Samek, W. Deep neural networks for no-reference and full-reference image quality assessment. IEEE Transactions on Image Processing 27(1), 206–219 (2018).

  66. 66.

    Ord, J. K. & Getis, A. Local spatial autocorrelation statistics: distribution issues and an application. Geog. Anal. 27(4), 286–306 (1995).

  67. 67.

    Bagstad, K. J., Semmens, D. J., Ancona, Z. H. & Sherrouse, B. C. Evaluating alternative methods for biophysical and cultural ecosystem services hotspot mapping in natural resource planning. Landscape ecology 32(1), 77–97 (2017).

  68. 68.

    Karl, J. W. & Maurer, B. A. Spatial dependency of predictions from image segmentation: a variogram-based method to determine appropriate scales for producing land-management information. Ecol. Inform. 5, 194–202 (2010).

  69. 69.

    Ming, J., Yang, L. & Li, Z. Song Modified ALV for selecting the optimal spatial resolution and its scale effect on image classification accuracy Math. Computer Model. 54(3–4), 1061–1068 (2011).

  70. 70.

    Kim, M. & Madden, T. Warner Estimation of optimal image object size for the segmentation of forest stands with multispectral IKONOS imagery In. T. Blaschke, S. Lang, G. J. Hay (Eds), Object-based Image Analysis–Spatial Concepts for Knowledge Driven Remote Sensing Applications, Springer, Berlin (2008).

  71. 71.

    Purser, A., Orejas, C., Gori, A., Unnithan, V. & Thomsen, L. Local variation in the distribution of benthic megafauna species associated with cold-water coral reefs on the Norwegian margin. Cont. Shelf Res. 54, 37–51 (2013).

Download references

Acknowledgements

The authors would like to thank Aurelian Arnaubec for his contribution to the video mosaicking data processing for this research. Authors would like to thank all cruise crew and scientific parties on RV Celtic Explorer (cruise number CE11009) with Holland 1 ROV. RV Celtic Explorer cruises were grant aided by the Marine Institute under the Ship Time Program of the National Development Plan, Ireland. This publication has emanated from research supported in part by a research grant from FAPESP (Fundação de Amparo a Pesquisa no Estado de São Paulo), Brazil under Grant 2017/19649-8. Science Foundation Ireland (SFI) under Grant Number 16/IA/4528 (co-funded by the Irish Marine Institute and the Geological Survey, Ireland) and by the Irish Research Council Graduate of Ireland Scholarship program (grant number GOIPG/2015/2700).

Author information

All authors conceived the presented idea. A.W. and A.L. were responsible for the fieldwork and data acquisition. L.C. developed the theory and performed the computations. All authors verified the analytical methods discussed the results and contributed to the final manuscript.

Correspondence to Luis A. Conti.

Ethics declarations

Competing Interests

The authors declare no competing interests.

Additional information

Publisher’s note: Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Further reading

Comments

By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.