Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Marine ship instance segmentation by deep neural networks using a global and local attention (GALA) mechanism

Abstract

Marine ships are the transport vehicle in the ocean and instance segmentation of marine ships is an accurate and efficient analysis approach to achieve a quantitative understanding of marine ships, for example, their relative locations to other ships or obstacles. This relative spatial information is crucial for developing unmanned ships to avoid crashing. Visible light imaging, e.g. using our smartphones, is an efficient way to obtain images of marine ships, however, so far there is a lack of suitable open-source visible light datasets of marine ships, which could potentially slow down the development of unmanned ships. To address the problem of insufficient datasets, here we built two instance segmentation visible light datasets of marine ships, MariBoats and MariBoatsSubclass, which could facilitate the current research on instance segmentation of marine ships. Moreover, we applied several existing instance segmentation algorithms based on neural networks to analyze our datasets, but their performances were not satisfactory. To improve the segmentation performance of the existing models on our datasets, we proposed a global and local attention mechanism for neural network models to retain both the global location and semantic information of marine ships, resulting in an average segmentation improvement by 4.3% in terms of mean average precision. Therefore, the presented new datasets and the new attention mechanism will greatly advance the marine ship relevant research and applications.

1 Introduction

Image segmentation plays an essential role in many visual understanding and object detection systems [13]. It involves a process that employs the intensity (brightness) or other information (e.g., edge) of an image to divide the image into independently connected regions. Image segmentation algorithms can be classified into at least two categories, i.e., semantic segmentation and instance segmentation. Semantic segmentation performs pixel-level labelling using a set of colors (object categories), while instance segmentation extends semantic segmentation by simultaneously detecting and delineating each object of interest in an image [3, 4]. Compared to object detection which merely detects the location of an object and places a window over it, instance segmentation performs more like a combination of object detection and semantic segmentation that not only detects the locations of all specific objects but also outlines and classifies individual detected objects [5]. As to marine ship segmentation, semantic segmentation classifies all ships in an image into one category, by labelling all ships with one color, while instance segmentation detects individual ships and classifies them into different categories. The applications of instance segmentation have been launched successfully in scenarios such as unmanned vehicle development [6, 7], human-computer interaction [8, 9], bio-medicine development [1012], video surveillance [1315], and marine ship monitoring [1618].

Since marine ships are the vehicle of ocean-related activities such as marine scientific research and education, transoceanic transport and marine fishing industry, image and video analysis of marine ships including instance segmentation of marine ships has received an increasing attention in the past years [1820]. Instance segmentation of marine ships is capable of providing important information such as the relative location of a ship with respect to other ships or surrounding obstacles, which is crucial for ship travel safety. Nowadays, various types of imaging techniques such as radar and infrared camera, have been equipped on a modern marine ship or an intelligent unmanned ship [16]. The spatial analysis of the ship imaging data can provide accurate environmental information to assist the ships to autonomously avoid crashing with other ships and natural obstacles in the ocean. Moreover, by segmenting a ship with respect to the background that may contain spatial information (for example, a known island or city), we can obtain the identity of the ship and the absolute location of the ship on earth. Therefore, instance segmentation of marine ships from a complex maritime background is essential for many ocean-related activities.

At present, satellite, synthetic aperture radar (SAR), infrared imaging (IR) and visible light (VL) imaging, are the main imaging tools to record marine ships, resulting in several different types of open-source databases for marine ship segmentation. These databases include satellite remote sensing images [2123], SAR images [24, 54], IR images [2527], and VL images [28]. The satellite images usually have a very large field of view, covering a wide space, but their image resolution is low, providing no accurate information (e.g. the shape and type) of a ship. SAR imaging can perform regardless of weather conditions, but SAR images usually contain a large amount of scattering noise and do not have rich spectral information, which is not convenient for subsequent ship segmentation purpose.IR imaging has strong penetration capability and is not easily affected by environmental conditions, but the contrast and signal-to-noise ratio of the obtained IR images are usually not high enough, resulting in a lack of color and texture information of ships. In contrast, VL images have many unique advantages over other types of images such as high resolution, the inclusion of color and texture information, high signal-to-noise ratio, high contrast and of rich details, and thus VL images can be a strong complementary portion to satellite, SAR and IR images [16]. With these merits, VL images can provide clear information of ship features (e.g., shape) that are crucial for subsequent ship detection, segmentation and classification. In addition, VL images can be easily obtained in a very low-cost way by using our routine cell phones and cameras, and therefore, they are suitable to build applications that need large scale data. However, so far, there are only a limited number of open-source databases for VL marine ship images. Although two VL image datasets of marine ships were presented by Zhang et al. and Sun et al. respectively [16, 18], these datasets have not been made publicly available, the scale of these datasets is relatively small and the labelling of the datasets in terms of ship categories is not fine enough.

The instance segmentation of marine ships in VL images is a challenging image processing task. In general, the existing segmentation methods can be mainly classified into thresholding [29], segmentation based on edge [30], region [31], super-pixel [32, 33], correlation theory [34, 35] and deep learning [36, 37]. The deep learning approach has gained increasing attention recently, and a challenge of instance segmentation based on deep learning is the acquisition of the location and semantic mask of each instance. Mask R-CNN implemented a general framework that can efficiently detect objects in an image while simultaneously generating a high-quality segmentation mask for each instance, which extends the Faster R-CNN model by adding a branch to predict an object mask in parallel with the existing branch for bounding box recognition [5]. Since the two-stage instance segmentation approach has high accuracy but suffers from low speed, the single-stage instance segmentation approach has been proposed to improve the segmentation efficiency. BlendMask was proposed to achieve improved mask prediction by using an effective combination of instance-level and semantic information with lower-level fine-granularity [36]. PolarMask was proposed to formulate an instance segmentation problem as the prediction of the instance contours through instance center classification and dense distance regression in a polar coordinate system, providing a new way of designing mask contours [38]. CenterMask is a single-stage anchor-free instance segmentation method that designed a new spatial attention-guided mask branching [37]. Different from the abovementioned approaches that rely on accurate edge detection, models such as SOLO and SOLOv2 can directly segment instance masks and learn instance mask labels, enabling end-to-end optimization [39, 40]. The SOLO algorithms were demonstrated that they could outperform both the two-stage and one-stage algorithms. Instance segmentation of marine ships in VL images has also attracted increasing attention in the field. Zhang et al proposed an integrated ship segmentation method based on discriminators and extractors to reduce the interference factors of complex ocean backgrounds [18]. To preserve the global information of a ship, Sun et al proposed a method using precise RoI pooling and global mask head that could improve the performance of ship instance segmentation [16]. showing that the use of the global and complete appearance information of a ship can increase the performance of instance segmentation. However, this method is too complex due to the extensive modification in the network architecture, and there is still some room for further improvement by the preservation of both the global and local ship information, while it is not trivial for the existing deep learning models to effectively retain both the global and local information.

The human attention mechanism can be a potential solution to retain both global and local information. In a complex scene, human attention can be attracted easily by salient features and regions. Inspired by this observation, the attention mechanism was introduced into computer vision. The use of an attention mechanism in instance segmentation can guide the segmentation to the most important regions of an image and ignore irrelevant parts [41]. The attention mechanism amplifies the role of key feature maps by assigning them greater weights. It is also important to note that the attention mechanism is a plug-and-play module that can be efficiently plugged into many deep learning models, which leads to a great success in the fields such as image classification, object detection, semantic segmentation, instance segmentation and 3D vision [4251]. In the existing deep neural networks, e.g. SENet and CBAM [42, 43], the attention mechanism was mainly used to convert the 2D feature maps into pixel feature maps by dimensionality reduction via 2D global pooling for feature map weight recalibration. These models using 2D global pooling mainly emphasized on the global information while ignored the local information. In contrast, the one-dimensional strip pooling is able to retain merely the local information along the spatial direction accurately [52]. Therefore, the combination of 2D global pooling and 1D strip pooling is a promising approach to preserve both the global and local information of an image.

In this work, in order to meet the urgent demand for open-source VL databases of marine ships, we collect and label two VL marine ship datasets, and test the segmentation performance of several existing deep learning models on our VL datasets. Moreover, we further propose a global and local attention (GALA) mechanism to improve the performance of the existing instance segmentation models, by combining 2D global pooling and 1D strip pooling to retrain both the global and local feature information. Both the datasets and the proposed approach are open-source available together with this study.

2 Methods

2.1 Analysis of the existing ship-related databases

To investigate whether the existing databases of marine ships contain sufficient VL images that can be used for marine ship instance segmentation, we first explored the open-source datasets, covering several image types, reported in previous studies (as shown in Table 1). These datasets included the VL dataset Sea Ships [53], the IR dataset Distant sea ships [54], the SAR dataset SAR-Ship-Dataset and SSDD [55, 56], the SAR dataset HRSID [57], and the MS COCO dataset that contains VL images of marine ships [58]. Sea Ships, which contains 31455 VL images and covers six commonly seen ship types, is mostly used for the object detection task. The Distant sea ships dataset consists of merely 3132 images of the long-wavelength IR type. The SAR-Ship-Dataset and SSDD datasets are only composed of high-resolution SAR images. In addition, these datasets are mainly designed for object detection purpose, by placing a window on a ship detected, and they do not have annotations required for instance segmentation purpose which needs to label the location, shape, and category of individual ships. HRSID can be used for ship instance segmentation because it contains the annotations required for instance segmentation, but it is of SAR type. The MS COCO dataset is a large open-source dataset commonly used for instance segmentation. To investigate whether this dataset contains specific VL images of marine ships that can be used for instance segmentation, we conducted a detailed analysis of the names of all images contained in the MS COCO dataset, which could indicate the contents or categories (person, car, boat, and so on) presented in the images. Fig 1 shows the number of individual categories in the MS COCO dataset, indicating that the distribution of the quantities of individual categories is uneven. The ‘person’ category takes up to 54% of the total number of images, while only about 2% of the images belongs to the ‘ship’ category.

thumbnail
Fig 1. Distribution of the numbers of individual categories in the MS COCO dataset.

The dataset has a total number of 80 categories, where the green and red arrows represent the number of images in the ‘person’ and ship category respectively.

https://doi.org/10.1371/journal.pone.0279248.g001

thumbnail
Table 1. Overview of several existing open-source datasets of marine ships.

https://doi.org/10.1371/journal.pone.0279248.t001

We then developed an image extraction script to extract the VL images of marine ships from the MS COCO dataset, and we named this subset as coco_boats. Since the image size and object scale in an image are important factors that can affect the performance of an algorithm, we also analyzed these two factors of images in coco_boats (Fig 2). We found that the length and width of the images are clustered at 500 and 650 pixels (red, Fig 2), which are evenly distributed in a line, while the length and width of the ships are located within [50, 650] and [50, 450] respectively (blue, Fig 2). This analysis shows that the image size and ship scale are not diverse enough. Taken together, IR images and SAR images took up most of the open-source marine ship datasets, while currently there is a lack of open-source VL datasets of marine ships that can be used for instance segmentation, and we extracted the coco_boats dataset from MS COCO to facilitate the research of marine ship instance segmentation.

thumbnail
Fig 2. The distribution of image and object sizes for the MS COCO ‘ship’ category.

A red triangle in the plot indicates the size of an image. A blue circle indicates the size of a ship.

https://doi.org/10.1371/journal.pone.0279248.g002

2.2 Two new datasets of VL images for marine ship instance segmentation

A reason of insufficient open-source VL datasets for marine ship instance segmentation may be the difficulty in data collection and the relatively time-consuming and laborious nature of labelling segmentation data. To overcome the problem of insufficient datasets, we developed a script to collect images from the ‘Google Image’ platform, using ‘ship’ as the searching keyword, after which we manually selected the true marine ship images and labelled the segmentations of ships using the LabelMe software [59]. Different from placing a window over a ship used for object detection purpose, which is commonly used in the object detection field, labelling ships for instance segmentation purpose is a very time-consuming task. In this labelling procedure, we first drew a polygon mask of a ship by following and marking the shape of the ship in an image without interruption, and then we named and classified the labelled ship. After delineating all the ships in an image with polygonal annotations, we generated an image annotation file using the json format, which will be later fed into a neural network. In total, we labelled a number of 6.2k images, which took about 400 hours. Note that the naming and annotation methods of the new datasets are in consistent with those of MS COCO.

From the Google Image platform, we generated two types of marine ship instance segmentation datasets, named as MariBoats and MariBoatsSubclass respectively, which we hoped can be used for different research purpose. The MariBoats dataset used all the 6.2k images and all the labelled ships were assigned to only one category, namely ‘ship’, resulting in 15.7k ship segmentation annotations. Compared with MariShipInsSeg [16], our dataset has a higher number of images and is open-source. This dataset with one category can satisfy the basic instance segmentation requirements (For example, avoiding obstacles (ships) during unmanned driving in the complex sea scene). To obtain a finer distinction of the labelled ships, we built MariBoatsSubclass, containing 3.1k images and 4.5k ship annotations. This dataset has six categories of marine ships: Engineering Ship (Eng.), Cargo Ship (Carg.), Speedboat (Sp.), Passenger Ship (Pass.), Official Ship (Off.), and Unknown Ship (Unk.). This dataset can be used for both the ship instance segmentation and the precise identification of marine ship categories in marine scenes.

2.2.1 MariBoats.

The MariBoats dataset is comprised of 6,271 ship images and 15,777 ship segmentation annotations, having only one category. The images in this dataset were partially extracted from 13717 ship images searched on ‘Google Image’ using keywords such as cargo ships, fishing boats, etc. We excluded those images of low quality, blurred, misrelated to ships, and the ones with duplicate content. We also included the coco_boats dataset (the subset of the MS COCO dataset containing VL images of ships) into MariBoats. Fig 3 shows our delineations of individual ships in representative images of MariBoats, illustrating the laborious nature of the delineation work. Avoiding the work of repeating such delineations is the motivation to make our datasets publicly available.

thumbnail
Fig 3. Representative images from MariBoats with manual annotations.

The manual delineations are indicated in blue, yellow or red.

https://doi.org/10.1371/journal.pone.0279248.g003

To distinguish the images collected from Google Image with coco_boats, we named the former as the ‘self_boats’ dataset. The distribution of the image size and ship scale of self_boats is shown in Fig 4. Compared with coco_boats (Fig 4), the image size distribution (black, Fig 2) of the self_boats dataset is located within [50, 800] in length and [50, 750] in width, and the size of ships (green, Fig 4) is distributed within [50, 550] in length and [50, 500] in width. From the scatter plots of image size and ship scale distribution of MariBoats, which is a combination of self_boats and coco_boats (Fig 5), we see that the image size and ship scale of the MariBoats dataset are more diverse than coco_boats.

thumbnail
Fig 4. The distribution of image and ship sizes in the self_boats dataset.

https://doi.org/10.1371/journal.pone.0279248.g004

thumbnail
Fig 5. The distribution of image and ship sizes in the MariBoats dataset.

It is a combination of Figs 2 and 4.

https://doi.org/10.1371/journal.pone.0279248.g005

2.2.2 MariBoatsSubclass.

To achieve a finer distinction of ship categories presented in individual images that can be used for ship instance segmentation, we constructed another dataset, namely MariBoatsSubclass, containing six categories as mentioned above. The dataset has 3125 high-quality VL images and 4588 labels. The number of labels are higher than that of VL images, which is because a single image may contain multiple ships of different types. Fig 6(a) shows the histogram of the number of ships in individual categories, among which the ‘Speedboat’ category has the highest number of images and labels (623 and 892, respectively). The category of ‘Unknown Ship’ has the lowest number of images, which is 469. In general, the distribution of each category and the accompanying segmented annotations are relatively evenly balanced, and the delineation of representatives of each category is shown in Fig 6(b).

thumbnail
Fig 6. The MariBoatsSubclass dataset.

(a) The histogram of the number of images and labelled ships in individual categories in MariBoatsSubclass. (b) Delineations of representative images of each category in MariBoatsSubclass.

https://doi.org/10.1371/journal.pone.0279248.g006

2.3. Global and local attention mechanism

After building two datasets for instance segmentation purpose, we next sought to test the performance of the existing instance segmentation models on our datasets. We tested classical models Mask R-CNN, SOLO and SOLOv2, and we found that the existing segmentation models to some extent cannot segment ships accurately because some ships were not detected by them (the red circle in Fig 7) [5, 39, 40]. The reason for the missed detection may due to the insufficient receptive field of the model, as discussed in [60], and the receptive field size of most CNN models is not proportional to its layer depth. An insufficient receptive field means that the global information of the inputs detected by a CNN model is not rich enough and one solution to this is to increase the receptive field. Pooling, a key element of CNN, is such a technique that can be adopted to increase the receptive field size of each convolutional kernel by downscaling the input graph, allowing the convolutional kernel to perceive a larger range of information from the input feature map.

thumbnail
Fig 7. An example of a missed detection by the existing algorithms.

The red circle represents the ship that should be detected but is not detected. Inset: the raw image.

https://doi.org/10.1371/journal.pone.0279248.g007

Pooling can be further divided into 2D spatial pooling and 1D strip pooling, and the difference between how they work in instance segmentation is illustrated in Fig 8 [52]. 2D pooling is more “global” as it downscales and converts the 2D feature maps of a certain length and width into individual pixels (Fig 8b). With such a downscaling, the convolutional kernel of a neural network can focus on the global feature information of the inputs. 1D strip pooling applies dimensionality reduction in only one direction, by projecting the 2D feature map along the vertical or horizontal direction into one dimension. As shown in Fig 8c, the direction in yellow indicates that dimensionality reduction is applied along the vertical direction, while the horizontal direction remains unchanged (red). Similarly, dimensionality reduction can be applied along the horizontal direction (red), while keeping the vertical direction unchanged. In this way, the information in the feature map can be preserved locally and thus the convolutional kernel of the neural network can perceive the local information of an object. Intuitively, when applied to marine ship images, global feature information represents the relative spatial relationships between different classes of objects [61], and for example, the pixels representing the sky and the sea are mostly above and below the pixels of ships, respectively. This global spatial intra-class correlation can be used to improve the performance of CNN models for ship instance segmentation. If we further consider the relative local information such as that both sky and ship are located above the sea but the ship tends to be closer to the sea, this local information allows us to model more complex spatial relationships [62]. If a CNN model only considers local feature information, suppose that the CNN model detects a ship, then adding the local feature will only motivate the CNN to further detect the ships nearby, while for the ships located far away, the local feature will not take effect. Therefore, inspired by [61], in order to obtain a better segmentation performance than the existing CNN instance segmentation models aforementioned, we aimed to combine the global and local information.

thumbnail
Fig 8. Illustration of the difference between 2D spatial pooling and 1D strip pooling in instance segmentation.

(a) The original image. (b) 2-D spatial pooling. (c)1D strip pooling. (d) The segmentation result using the GALA mechanism.

https://doi.org/10.1371/journal.pone.0279248.g008

To achieve this combination, we noticed that the global pooling mechanism is often implemented with the attention mechanism, an area under active investigation currently [43, 63], and we also noticed that these attention mechanism models mostly do not use the local feature information. In [48], a model using local pooling was proposed, but it did not use the global feature information. Here, in order to benefit from both the global and local pooling mechanism [61], we implemented the combination of the global and local mechanism together with the attention mechanism. By using both the global 2D spatial pooling and the local 1D strip pooling, we hoped that the new GALA mechanism is capable of improving segmentation performance. An example of the segmentation results by using the proposed GALA mechanism is shown in Fig 8d.

The GALA mechanism was implemented as follows, schematically shown in Fig 9. First, we used the global average pooling to reduce the dimensionality of the entire 2D spatial information in a feature map, which is averaging all pixel values of each channel map, and we obtained a new 1 × 1 channel map. The output of the c-th channel feature map can be expressed as [43], (1) where xc is the input image of the c-th channel, H is the height of the input image, W is the width of the input image and zc is the output image of the c-th channel.

thumbnail
Fig 9. The scheme depicts the GALA mechanism.

C represents the number of channels, that is, the number of feature maps.H and W represents the height and width of the feature maps, respectively.

https://doi.org/10.1371/journal.pone.0279248.g009

After processed by the activation function and convolutional transformation, the feature map of channel correlation can be obtained and the output image of the c-th channel can be expressed as, (2) where σ denotes the sigmoid function, and T1 and T2 are nonlinear transformations describing the importance of each channel. δ denotes the ReLU activation function. Second, we performed 1D pooling of the feature map with channel correlation in the horizontal and vertical direction respectively, and the output image of the c-th channel with height h can be expressed as, (3)

The output image of channel c with width w can be expressed as, (4) where denotes the input image of the c-th channel that already has channel correlation. After processing by the activation function and convolutional transformation, the final feature maps of channel correlation, direct perception, and position sensitivity can be obtained. The output image of the c-th channel can be expressed as, (5) where Fh and Fw denote the two 1 × 1 convolutional transforms.

In order to determine which layers of a neural network the GALA mechanism is specifically applied to, we deployed GALA into a feature pyramid network containing multi-scale feature information. The shallow feature maps of the feature pyramid network characterize the detailed information of an object and the deep feature maps characterize the semantic information of an object. The combination of feature maps with different depths within the network forms multi-scale representation information. To take full advantage of the multi-scale representation in the feature pyramid network, here we propose an Enhanced Feature Pyramid Network (EFPN) based on the GALA mechanism, as shown in Fig 10. For each layer of the different scale predictions of the feature pyramid network, the GALA mechanism is used to enhance the representation of the feature graph.

thumbnail
Fig 10. Schematic diagram of the enhanced feature pyramid structure.

EFPN represents our enhanced improvement of the feature pyramid. GALA represents our proposed global and local attention mechanism. C1-C4 represent the feature maps of different sizes extracted from the backbone network, respectively, and P1-P4 represent the predicted objects of different scales, respectively. The original structure of the feature pyramid network is shown in the lower right.

https://doi.org/10.1371/journal.pone.0279248.g010

3 Experiment results

Transfer learning is a technique to avoid training a neural network from scratch, and here we used the pre-trained ImageNet model as the start point for retraining using our datasets [64]. The experimental environment was configured with Ubuntu 16.04.4, pytorch1.4.0, and 4 NVIDIA GeForce GTX 1080Ti GPUs. The learning rate is 0.01. The optimization algorithm is the Stochastic Gradient Descent. We first tested the performance of several classical instance segmentation models, namely Mask R-CNN, SOLO and SOLOv2 models, on our datasets which included coco_boats, self_boats, MariBoats (coco_boats + self_boats), and MariBoatsSubclass. The datasets and accompanying algorithms can be download via the link, https://github.com/s2120200252/Visible-ship-dataset.

3.1 Performance evaluation metrics

We followed the quantitative metric system used by the MS COCO dataset to evaluate the performance of ship instance segmentation models. The six metrics used are Intersection of Union (IoU), Average Precision (AP), Mean Average Precision (mAP), Frames Per Second (FPS), Parameter (Para.) and Time Complexity (TC). We refer to a more detailed definition about these metrics [58]. IoU is defined as the degree of overlap between two segmentations. AP is the major metric that will be used for accuracy determination. (6) where P represents precision and R represents recall rate. AP calculates an average IoU, averaging from 0.5 to 0.95 with increment of 0.05. For example, AP50 and AP75 represent the calculation of the average IoU at thresholds of 0.5 and 0.75, respectively. For multi-scale object detection capability, APS, APM, and APL are used to represent the average accuracy of objects of small (area < 322 pixels), medium (322 < area < 642 pixels), and large (area > 642 pixels) size, respectively. Mean Average Precision(mAP) is a mean value of AP over the number of categories of ships to be detected. In addition to evaluate the accuracy, it is also important to compare the running time of the tested instance segmentation models. FPS is such a metric that measures the number of images a model can process in a second. Parameter is the number of parameters in a neural network model (the parameters learned when training the network). (7) where Co represents the number of input channels, k represents the size of a convolution kernel, and Ci represents the number of output channels. Generally speaking, the number of parameters is positively proportional to the memory required to save the model and the hardware memory requirement. The metric Memory Access Cost (MAC) is generally used to measure the time complexity (TC) of a model or an algorithm, defined as, (8) where Co represents the number of input channels, k represents the size of a convolution kernel, and Ci represents the number of output channels, W represents the width of a feature map, H represents the height of a feature map.

3.2 Performance test

3.2.1 Segmentation results on MariBoats having one ship category.

We next moved onto the performance evaluation of the existing models, i.e. Mask R-CNN, SOLO and SOLOv2, on our datasets. We first performed the evaluation on the coco_boats dataset, and we found that the segmentations of the images in coco_boats by all the three models were not satisfactory (AP: 9.4,13.4,13.6), which are mostly due to the fact that these models have not been trained on enough VL ship images. Actually, the poor performance of these models on coco_boats was our initial motivation for building a larger dataset, containing sufficient VL images of marine ships, and we hoped that retraining the models by this larger dataset, MariBoats, can improve the performance of these models. We then used 70% of MariBoats to retrain the three models aforementioned, and used the remaining 30% for testing. The results of the quantitative tests are shown in Tables 24, which directly indicate that the segmentation accuracy of the three network models has improves for both MariBoats and the two subsets of MariBoats, coco_boats and self_boats, after retraining the models by MariBoats (the last row in Tables 24).

thumbnail
Table 2. Test results of Mask-RCNN in three instance segmentation datasets.

https://doi.org/10.1371/journal.pone.0279248.t002

thumbnail
Table 3. Test results of SOLO in the three instance segmentation datasets.

https://doi.org/10.1371/journal.pone.0279248.t003

thumbnail
Table 4. Test results of SOLOv2 in the three instance segmentation datasets.

https://doi.org/10.1371/journal.pone.0279248.t004

Meanwhile, we further tested the segmentation accuracy of the three models by using only coco_boats or self_boats to retrain the models and using the other two datasets for testing (row 1-2 in Tables 24). Taking the SOLOv2 as an example (Table 4), the segmentation accuracy of coco_boats shows that the model trained on the MariBoats training set improves by 2.4% compared with the results trained on the coco_boats training set. The results of self_boats show that the model trained on the MariBoats training set improves by 5.7% compared with the results trained on the self_boats training set. The MariBoats results show that the model trained on the MariBoats training set improves by 7.9% compared with the results trained on the self_boats training set. Tables 2 and 3 further verify the advantages of the MariBoats dataset, containing richer image data, in improving the segmentation performance of Mask R-CNN and SOLO, respectively. Taken together, these data fully validated the necessity of building the self_boats dataset and the MariBoats dataset.

3.2.2 Segmentation results on MariBoatsSubclass having six ship categories.

We continued to test the segmentation performance and computational speed of Mask R-CNN, SOLO, and SOLOv2 on MariBoatsSubclass upon solving an instance segmentation problem having six ship categories. The experimental results are shown in Table 5. We first set the backbone network ResNet to commonly used 50 layers for Mask R-CNN, SOLO, and SOLOv2 (row 1-3 in Table 5), and SOLOv2 had the highest score in terms of mAP (57.8%), compared with 42.9% and 55.5% for Mask R-CNN and SOLO, respectively. The segmentation accuracy of these models on segmenting the Unknown ship category is generally much lower compared with other categories, and we reasoned that the smaller size of the ships (having fewer pixels) shown in this category can affect the evaluation of the IoU metric. However, the mAP of all six-class ships is higher. These further validated the diversity of ship scales in MariBoatsSubclass. We also noticed that although these model showed improvement after retraining by MariBoatsSubclass, there is still quite some room to improve the performance of these models. For example, these model are still suffering from incomplete segmentation, as shown in Fig 11.

thumbnail
Fig 11. A representative example of incomplete segmentation by Mask R-CNN, SOLO and SOLOv2.

The red circles represent the parts of ships that should be detected but are not detected. Insets: amplified images of the objects with the red circles.

https://doi.org/10.1371/journal.pone.0279248.g011

thumbnail
Table 5. Test results of different network structures on MariBoatsSubclass.

https://doi.org/10.1371/journal.pone.0279248.t005

Since SOLOv2 had the best performance, and we then selected SOLOv2 to test the computational speed using a different number of layers (row 4-5). The segmentation accuracies for SOLOv2 were 56.2% and 57.4% in terms of mAP when the backbone network ResNet was set to 18 and 34 layers, respectively. Although the mAP values decreased by 1.6% and 0.4%, respectively, the computational speed of SOLOv2 improved by 35% and 20% in terms of FPS, reaching 44.7 and 39.9, respectively. This suggests that the segmentation speed can be significantly improved with a small loss of AP by reducing the number of residual layers of the backbone network.

3.3 Performance testing of the proposed GALA mechanism

Based on the segmentation results of the MariBoatsSubclass dataset, SOLOv2 achieved the highest segmentation accuracy and was selected as the base network to implement the attention mechanism. Note that the attention mechanism has not been used in the existing models including Mask R-CNN, SOLO, and SOLOv2. Here, we not only introduced the attention mechanism into SOLOv2, and we further proposed a new attention mechanism, GALA, to improve the performance of the existing attention mechanisms. To prove the performance of the proposed GALA mechanism, we introduced different classical attention mechanisms to SOLOv2 as a comparison (Table 6). As shown in Table 6, compared with SOLOv2 without using any attention mechanism, the performance of all SOLOv2 models using several classical attention mechanisms ECA-Net [65], ScSE [63], Triplet Attention [66], SENet [43], and CA [48], improved in terms of mAP, and the one using GALA mechanism had the largest improvement, reaching 62.1%, which is 4.3% improvement with respect to SOLOv2 and 3.8%, 3.6% and 3.3% improvements with respect to CA, Triplet Attention and SEnet, respectively. This validated the performance of the proposed GALA attention mechanism over the existing attention mechanism, on the VL ship datasets.

thumbnail
Table 6. The performance of different attention mechanisms on MariBoatsSubclass.

https://doi.org/10.1371/journal.pone.0279248.t006

We further analyzed the computational complexity of the SOLOv2 model using GALA, in terms of the increase in the number of parameters and FPS. The increase in the number of parameters with respect to SOLOv2 was only 0.02 M. This increase is nearly negligible compared with the 30.94 M parameters of SOLOv2. The increase in time complexity was only 0.03 GMAC, and the FPS decreased by only 1.6. This analysis indicates that introducing GALA into the feature pyramid network can improve the performance of instance segmentation models significantly with little increase in time complexity and number of parameters.

Visually, GALA also improved the segmentation performance of SOLOv2 on MariBoats. As shown in Fig 12, SOLOv2, SOLOv2+SEnet, and SOLOv2+CA networks cannot detect completely all the ships, by either missing a ship partially (the red circles in Fig 12a–12c) or entirely (Fig 12i and 12k). In addition, SOLOv2, SOLOv2+SEnet, and SOLOv2+CA networks incorrectly detected the aircraft as a part of the ship below it (Fig 12e–12g). However, SOLOv2+GALA was able to correctly separate the aircraft from the ship, and could detect all the ships completely and correctly (Fig 12d, 12l, and 12h). Therefore, this visual comparison together with the quantitative comparison verified the superiority of the proposed GALA mechanism over the existing attention mechanisms.

thumbnail
Fig 12. Comparison of segmentation results for several typical ship examples, selected from MariBoatsSubclass.

The first row (a, e, i): segmentations of SOLOv2. The second row (b, f, j): the segmentations of SOLOv2+ SEnet. The third row (c, g, j): the segmentations of SOLOv2+ CA. The last row (d, h, l): the segmentations of SOLOv2 using the proposed GALA mechanism.

https://doi.org/10.1371/journal.pone.0279248.g012

4 Conclusion and discussion

Marine ship instance segmentation of VL images plays an important role in marine-related scientific research, educational and commercial applications. However, there are hardly any publicly available and suitable datasets containing VL images of marine ships for ship instance segmentation purpose. To address this, we collected and manually labelled two new VL marine ship datasets using a data collection tool that we developed. The datasets and the accompanying image processing tools are available to the public for visual perception applications of marine scenes. To the best of our knowledge, there are also rare instance segmentation methods that are specially designed for marine ship segmentation. Therefore, considering the special characteristics of marine ship instance segmentation, we proposed the GALA attention mechanism, which takes advantage of 1D strip pooling and 2D spatial pooling to preserve both the global and local information of the input images. Experimental results demonstrated the superiority of GALA over the existing attention mechanisms on the marine ship datasets. We selected Mask R-CNN, SOLO, and SOLOv2 as comparison because these models have been well recognized in the computer vision and remote sensing fields. Many works have tried to improve these models, such as SOLO series models [39, 40], and the Cascade R-CNN models [67], and these variants are also widely applied in many fields.

Sun et al. has built a VL dataset, MariShipInsSeg, which contains 4K images but MariShipInsSeg is not open-source and this dataset has one general ship category [16]. In contrast, our dataset MariBoats has more images and is open-source. In addition, we have further refined the category of MariBoats into six ship categories, leading to our second dataset, MariBoatsSubclass. It will facilitate the research on accurate segmentation and classification of marine ships based on appearance, shape and function. Moreover, we have introduced the attention mechanism into the marine ship instance segmentation of VL images and we have further proposed a novel attention mechanism which can improve the performance of existing neural networks. It is important to notice that although the attention mechanism has been applied to the COCO dataset before, there are no such attention models specially designed for marine ship instance segmentation of VL images. Marine ship images are different from most of the images in the COCO dataset in terms of background, texture, and contour of objects. Therefore, it is the first time that the attention mechanism is introduced into marine ship instance segmentation using VL images. Most of the existing attention mechanisms such as ECA-Net [65], ScSE [63], Triplet Attention [66], and SENet [43], have focused on the global information, and only a few attention mechanisms such as CA [48], have paid attention to the local information. Inspired by [61], we combined both the global and local mechanism to retain both the global and local feature information, achieving better segmentation results than the existing attention mechanisms using only global or local information alone. Our proposed GALA mechanism maintains the convenience of the attention mechanism that can be applied generally to most of the neural networks, and the source codes are immediately available to the public. Finally, we have also made the accompanying data extraction and analysis tools publicly available to facilitate research in the computer vision field.

In conclusion, we believe that the new open-source datasets we built in this work and the proposed GALA mechanism will facilitate research in VL ship applications and attract attention from other computer vision fields to use the GALA mechanism. Future work includes enriching the current datasets for marine ship instance segmentation and developing fast segmentation methods for segmenting ships from complex ocean scenes. Moreover, we also believe that the proposed mechanism is applicable to other fields such as remote sensing and biomedical microscopic data, because the instant segmentation of targets in these datasets, for example marine ships in SAR image datasets [57], and cells in the microscopic datasets [68, 69], also relies on the use of global and local information to describe the targets and to distinguish the targets from the backgrounds. The application of the GALA attention mechanism to these datasets will be a valuable plan for future to explore.

References

  1. 1. Gonzalez RC. Digital image processing: Pearson education india; 2009:368–369.
  2. 2. Zhang Z, Kuzmin NV, Groot ML, de Munck JCJB. Extracting morphologies from third harmonic generation images of structurally normal human brain tissue. Bioinformatics. 2017;33(11):1712–20. pmid:28130231
  3. 3. Minaee S, Boykov YY, Porikli F, Plaza AJ, Kehtarnavaz N, Terzopoulos DJItopa, et al. Image segmentation using deep learning: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2021;44(7):3523–3542.
  4. 4. Wang Q, Gao J, Li XJIToIP. Weakly supervised adversarial domain adaptation for semantic segmentation in urban scenes. IEEE Transactions on Image Processing. 2019;28(9):4376–86. pmid:30998470
  5. 5. He K, Gkioxari G, Dollar P, Girshick R. Mask R-CNN. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2020;42(2):386–97. pmid:29994331
  6. 6. Hu J-F, Sun J, Lin Z, Lai J-H, Zeng W, Zheng W-S. APANet: Auto-Path Aggregation for Future Instance Segmentation Prediction. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2022;44(7):3386–403. pmid:33571087
  7. 7. Liu S, Qi L, Qin H, Shi J, Jia J, Ieee, editors. Path Aggregation Network for Instance Segmentation. 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2018:8759-5768.
  8. 8. Liu H, Zhou A, Dong Z, Sun Y, Zhang J, Liu L, et al. M-Gesture: Person-Independent Real-Time In-Air Gesture Recognition Using Commodity Millimeter Wave Radar. IEEE Internet of Things Journal. 2022;9(5):3397–415.
  9. 9. Zhang H, Luo G, Tian Y, Wang K, He H, Wang F-Y. A Virtual-Real Interaction Approach to Object Instance Segmentation in Traffic Scenes. IEEE Transactions on Intelligent Transportation Systems. 2021;22(2):863–75.
  10. 10. Gurcan MN, Boucheron LE, Can A, Madabhushi A, Rajpoot NM, Yener B. Histopathological image analysis: a review. IEEE reviews in biomedical engineering. 2009;2:147–147. pmid:20671804
  11. 11. Liang P, Zhang Y, Ding Y, Chen J, Madukoma CS, Weninger T, et al. H-EMD: A Hierarchical Earth Mover’s Distance Method for Instance Segmentation. IEEE transactions on medical imaging. 2022;PP. pmid:35446762
  12. 12. Lienkamp SS, Liu K, Karner CM, Carroll TJ, Ronneberger O, Wallingford JB, et al. Vertebrate kidney tubules elongate using a planar cell polarity-dependent, rosette-based mechanism of convergent extension. Nature Genetics. 2012;44(12):1382–7. pmid:23143599
  13. 13. Dong M, Wang J, Huang Y, Yu D, Su K, Zhou K, et al., editors. Temporal Feature Augmented Network for Video Instance Segmentation. IEEE/CVF International Conference on Computer Vision (ICCV); 2019:721–721.
  14. 14. Wang Y, Xu Z, Wang X, Shen C, Cheng B, Shen H, et al., editors. End-to-End Video Instance Segmentation with Transformers. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2021:8737–8737.
  15. 15. Yang L, Fan Y, Xu N, Ieee, editors. Video Instance Segmentation. IEEE/CVF International Conference on Computer Vision (ICCV); 20195187-5196.
  16. 16. Sun Y, Su L, Luo Y, Meng H, Li W, Zhang Z, et al. Global Mask R-CNN for marine ship instance segmentation. Neurocomputing. 2022;480:257–257.
  17. 17. Wu Z, Hou B, Ren B, Ren Z, Wang S, Jiao L. A Deep Detection Network Based on Interaction of Instance Segmentation and Object Detection for SAR Images. Remote Sensing. 2021;13(13).
  18. 18. Zhang W, He X, Li W, Zhang Z, Luo Y, Su L, et al. An integrated ship segmentation method based on discriminator and extractor. Image and Vision Computing. 2020;93.
  19. 19. Fan F, Zeng X, Wei S, Zhang H, Tang D, Shi J, et al. Efficient Instance Segmentation Paradigm for Interpreting SAR and Optical Images. Remote Sensing. 2022;14(3).
  20. 20. Zeng X, Wei S, Wei J, Zhou Z, Shi J, Zhang X, et al. CPISNet: Delving into Consistent Proposals of Instance Segmentation Network for High-Resolution Aerial Images. Remote Sensing. 2021;13(14).
  21. 21. Cheng D, Meng G, Xiang S, Pan C. Efficient sea-land segmentation using seeds learning arid edge directed graph cut. Neurocomputing. 2016;207:36–36.
  22. 22. Huang G, Wan Z, Liu X, Hui J, Wang Z, Zhang Z. Ship detection based on squeeze excitation skip-connection path networks for optical remote sensing images. Neurocomputing. 2019;332:215–215.
  23. 23. Xu J, Sun X, Zhang D, Fu K. Automatic Detection of Inshore Ships in High-Resolution Remote Sensing Images Using Robust Invariant Generalized Hough Transform. IEEE Geoscience and Remote Sensing Letters. 2014;11(12):2070–4.
  24. 24. Ouchi K, Tamaki S, Yaguchi H, Iehara M. Ship Detection Based on Coherence Images Derived From Cross Correlation of Multilook SAR Images. IEEE Geoscience and Remote Sensing Letters. 2004;1(3):184–7.
  25. 25. Bai X, Liu M, Wang T, Chen Z, Wang P, Zhang Y. Feature based fuzzy inference system for segmentation of low-contrast infrared ship images. Applied Soft Computing. 2016;46:128–128.
  26. 26. Liu Z, Zhou F, Ieee Bai X, editors. INFRARED SHIP TARGET SEGMENTATION BASED ON REGION AND SHAPE FEATURES. 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS); 2013.
  27. 27. Han J, Liang K, Zhou B, Zhu X, Zhao J, Zhao L. Infrared Small Target Detection Utilizing the Multiscale Relative Local Contrast Measure. IEEE Geoscience and Remote Sensing Letters. 2018;15(4):612–6.
  28. 28. Chen X, Chen H, Wu H, Huang Y, Yang Y, Zhang W, et al. Robust Visual Ship Tracking with an Ensemble Framework via Multi-View Learning and Wavelet Filter. Sensors. 2020;20(3). pmid:32050581
  29. 29. Oliva D, Cuevas E, Pajares G, Zaldivar D, Perez-Cisneros M. Multilevel Thresholding Segmentation Based on Harmony Search Optimization. Journal of Applied Mathematics. 2013.
  30. 30. Senthilkumaran N, Rajesh R, editors. Image segmentation-a survey of soft computing approaches. 2009 International Conference on Advances in Recent Technologies in Communication and Computing; 2009: IEEE.
  31. 31. Ning J, Zhang L, Zhang D, Wu C. Interactive image segmentation by maximal similarity based region merging. Pattern Recognition. 2010;43(2):445–56.
  32. 32. Van den Bergh M, Boix X, Roig G, Van Gool L. SEEDS: Superpixels Extracted Via Energy-Driven Sampling. International Journal of Computer Vision. 2015;111(3):298–314.
  33. 33. Veksler O, Boykov Y, Mehrani P, editors. Superpixels and Supervoxels in an Energy Optimization Framework. 11th European Conference on Computer Vision; 2010.
  34. 34. Khokher MR, Ghafoor A, Siddiqui AMJIip. Image segmentation using multilevel graph cuts and graph development using fuzzy rule-based system. IET image processing. 2013;7(3):201–11.
  35. 35. Zhao Z, Cheng L, Cheng G. Neighbourhood weighted fuzzy c-means clustering algorithm for image segmentation. IET Image Processing. 2014;8(3):150–61.
  36. 36. Chen H, Sun K, Tian Z, Shen C, Huang Y, Yan Y, editors. Blendmask: Top-down meets bottom-up for instance segmentation. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2020.
  37. 37. Lee Y, Park J, editors. Centermask: Real-time anchor-free instance segmentation. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2020.
  38. 38. Xie E, Sun P, Song X, Wang W, Liu X, Liang D, et al., editors. Polarmask: Single shot instance segmentation with polar representation. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition; 2020.
  39. 39. Wang X, Zhang R, Shen C, Kong T, Li L. SOLO: A Simple Framework for Instance Segmentation. IEEE transactions on pattern analysis and machine intelligence. 2021;PP.
  40. 40. Wang X, Zhang R, Kong T, Li L, Shen CJae-p. SOlOv2: Dynamic, Faster and Stronger. arXiv e-prints. 2020:arXiv: 2003.10152.
  41. 41. Guo M-H, Xu T-X, Liu J-J, Liu Z-N, Jiang P-T, Mu T-J, et al. Attention mechanisms in computer vision: A survey. Computational Visual Media. 2022;8(3):331–68.
  42. 42. Woo S, Park J, Lee J-Y, Kweon IS, editors. CBAM: Convolutional Block Attention Module. 15th European Conference on Computer Vision (ECCV); 2018:3–19.
  43. 43. Hu J, Shen L, Albanie S, Sun G, Wu E. Squeeze-and-Excitation Networks. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2020;42(8):2011–23. pmid:31034408
  44. 44. Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S, editors. End-to-end object detection with transformers. European Conference on Computer Vision; Springer,2020:213–229.
  45. 45. Dai J, Qi H, Xiong Y, Li Y, Zhang G, Hu H, et al., editors. Deformable Convolutional Networks. 16th IEEE International Conference on Computer Vision (ICCV); 2017:764–773.
  46. 46. Fu J, Liu J, Tian H, Li Y, Bao Y, Fang Z, et al., editors. Dual Attention Network for Scene Segmentation. 32nd IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2019:3146–3154.
  47. 47. Yuan Y, Huang L, Guo J, Zhang C, Chen X, Wang JJapa. Ocnet: Object context network for scene parsing. arXiv preprint arXiv:00916. 2018.
  48. 48. Hou Q, Zhou D, Feng J, Ieee Comp SOC, editors. Coordinate Attention for Efficient Mobile Network Design. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2021.
  49. 49. Zhang H, Zu K, Lu J, Zou Y, Meng DJapa. EPSANet: An Efficient Pyramid Squeeze Attention Block on Convolutional Neural Network. arXiv preprint arXiv:2105.14447, 2021.
  50. 50. Guo M-H, Cai J-X, Liu Z-N, Mu T-J, Martin RR, Hu S-M. PCT: Point cloud transformer. Computational Visual Media. 2021;7(2):187–99.
  51. 51. Xie S, Liu S, Chen Z, Tu Z, Ieee, editors. Attentional ShapeContextNet for Point Cloud Recognition. 31st IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2018:4606–4615.
  52. 52. Hou Q, Zhang L, Cheng M-M, Feng J, Ieee, editors. Strip Pooling: Rethinking Spatial Pooling for Scene Parsing. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2020.
  53. 53. Shao Z, Wu W, Wang Z, Du W, Li C. SeaShips: A Large-Scale Precisely Annotated Dataset for Ship Detection. IEEE Transactions on Multimedia. 2018;20(10):2593–604.
  54. 54. X. Zhao ZL, Y. Li, S. Fan, J. Liu, L. Wang, J. Kang, et al. Distant sea (10-12km) ships. Available from: url: http://www.gxzx.sdu.edu.cn/info/1133/2174.htm,. 2020.
  55. 55. Wang Y, Wang C, Zhang H, Dong Y, Wei S. A SAR Dataset of Ship Detection for Deep Learning under Complex Backgrounds. Remote Sensing. 2019;11(7).
  56. 56. Li J, Qu C, Shao J, Ieee, editors. SHIP DETECTION IN SAR IMAGES BASED ON AN IMPROVED FASTER R-CNN. Conference on SAR in Big Data Era—Models, Methods and Applications (BIGSARDATA); 2017:1–16.
  57. 57. Wei S, Zeng X, Qu Q, Wang M, Su H, Shi J. HRSID: A High-Resolution SAR Images Dataset for Ship Detection and Instance Segmentation. IEEE Access. 2020;8:120234–120234.
  58. 58. Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, et al., editors. Microsoft COCO: Common Objects in Context. 13th European Conference on Computer Vision (ECCV); 2014,740-755.
  59. 59. Russell BC, Torralba A, Murphy KP, Freeman WT. LabelMe: A database and web-based tool for image annotation. International Journal of Computer Vision. 2008;77(1-3):157–73.
  60. 60. Zhao H, Shi J, Qi X, Wang X, Jia J, Ieee, editors. Pyramid Scene Parsing Network. 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); 2017:2881–2890.
  61. 61. Gould S, Rodgers J, Cohen D, Elidan G, Koller D. Multi-class segmentation with relative location prior. International Journal of Computer Vision. 2008;80(3):300–16.
  62. 62. Fink M, Perona P, editors. Mutual boosting for contextual inference. 17th Annual Conference on Neural Information Processing Systems (NIPS); 2003:16.
  63. 63. Roy AG, Navab N, Wachinger C, editors. Concurrent Spatial and Channel’Squeeze & Excitation’ in Fully Convolutional Networks. 21st International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI) 8th Eurographics Workshop on Visual Computing for Biology and Medicine (VCBM) International Workshop on Computational Diffusion MRI (CDMRI); 2018:421-429.
  64. 64. He K, Girshick R, Dollár P, editors. Rethinking imagenet pre-training. Proceedings of the IEEE/CVF International Conference on Computer Vision; 2019:4918-4927.
  65. 65. Wang Q WB, Zhu P, et al.. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). 2020:13-19.
  66. 66. Misra D, Nalamada T, Arasanipalai AU, Hou Q, Ieee, editors. Rotate to Attend: Convolutional Triplet Attention Module. IEEE Winter Conference on Applications of Computer Vision (WACV); 2021:3139-3148.
  67. 67. Cai Z.; Vasconcelos N. Cascade R-CNN: High Quality Object Detection and Instance Segmentation. Ieee Transactions on Pattern Analysis and Machine Intelligence 2021, 43, 1483–1498. pmid:31794388
  68. 68. Zhang Z.; Kuzmin N.V.; Groot M.L.; de Munck J.C. Quantitative comparison of 3D third harmonic generation and fluorescence microscopy images. Journal of Biophotonics 2018, 11. pmid:28464543
  69. 69. Zhang Z.; de Munck J.C.; Verburg N.; Rozemuller A.J.; Vreuls W.; Cakmak P.; et al. Quantitative Third Harmonic Generation Microscopy for Assessment of Glioma in Human Brain Tissue. Advanced Science 2019, 6. pmid:31179222