Ucloe 05 058
Ucloe 05 058
                                          How to cite
                                          Jialeng G, Fuente S, Smith T. BoatNet: automated small boat composition detection using deep learning on
                                          satellite imagery. UCL Open: Environment. 2023;(5):05. Available from: https://doi.org/10.14324/111.444/
                                          ucloe.000058
Submission date: 20 July 2022; Acceptance date: 3 April 2023; Publication date: 24 May 2023
                                          Peer review
                                          UCL Open: Environment is an open scholarship publication, this article has been peer-reviewed through the
                                          journal’s standard open peer review process. All previous versions of this article and open peer review reports
                                          can be found online in the UCL Open: Environment Preprint server at ucl.scienceopen.com
                                          Open access
                                          This is an open access article distributed under the terms of the Creative Commons Attribution Licence
                                          (CC BY) 4.0 https://creativecommons.org/licenses/by/4.0/, which permits unrestricted use, distribution and
                                          reproduction in any medium, provided the original author and source are credited.
*Corresponding author
                                          Abstract
E-mail: jialeng.guo@ucl.ac.uk             Tracking and measuring national carbon footprints is key to achieving the ambitious goals set
1
 UCL Energy Institute, The Bartlett       by the Paris Agreement on carbon emissions. According to statistics, more than 10% of global
School of Environment, Energy             transportation carbon emissions result from shipping. However, accurate tracking of the emissions
and Resources, University College         of the small boat segment is not well established. Past research looked into the role played by
London, London, UK                        small boat fleets in terms of greenhouse gases, but this has relied either on high-level technological
                                          and operational assumptions or the installation of global navigation satellite system sensors to
                                          understand how this vessel class behaves. This research is undertaken mainly in relation to fishing
                                          and recreational boats. With the advent of open-access satellite imagery and its ever-increasing
                                          resolution, it can support innovative methodologies that could eventually lead to the quantification
                                          of greenhouse gas emissions. Our work used deep learning algorithms to detect small boats in
                                          three cities in the Gulf of California in Mexico. The work produced a methodology named BoatNet
                                          that can detect, measure and classify small boats with leisure boats and fishing boats even under
                                          low-resolution and blurry satellite images, achieving an accuracy of 93.9% with a precision of
                                          74.0%. Future work should focus on attributing a boat activity to fuel consumption and operational
                                          profile to estimate small boat greenhouse gas emissions in any given region.
Keywords: object detection, deep learning, machine learning, transfer learning, small boat activity, climate change
1 / 20                                BoatNet: automated small boat composition detection using deep learning on satellite imagery     UCL OPEN ENVIRONMENT
                                      https://doi.org/10.14324/111.444/ucloe.000058
                                      BoatNet: automated small boat composition detection using deep learning on satellite imagery
         Introduction
         Energy crisis, energy security and climate change
         The Intergovernmental Panel on Climate Change (IPCC) explains, in its latest report, that humans
         and nature are being pushed beyond their abilities to adapt due to the anthropogenic emissions
         caused by economic, industrial and societal activities [1]. Carbon-intensive resources still comprise
         a large proportion of the energy system [1] – about 80% in 2017 [2]. However, the share of
         electricity production from renewables increased from 20.8% to 29.0% between 1985 and 2020 [3].
         Still, carbon emissions have not been reduced in line with the ambitions of the Paris Agreement,
         and it is predicted that in the next few years, the gains in carbon reduction due to the Covid-19
         pandemic will be erased, faster than expected [4]. However, even under all these pressures and
         projections, it is still possible for humanity to keep the global temperature below 1.5°C from pre-
         industrial levels by 2100 if substantial changes are made to the current energy systems.
         However, energy security is an important part of the strategies proposed by countries to support
         economic growth and provide essential services to their populations. Currently, nations deposit
         most of their energy security into fossil fuels while expanding their renewable power capacity. Fossil
         fuels and their conversion systems (e.g., internal combustion engines) permit operators to react
         quickly to changes in the energy demand (i.e., more control over energy deployment) while offering
         acceptable volumetric energy densities. However, heavy reliance on fossil fuels, coupled with the
         fuel’s geographical origin, is at the mercy of important price fluctuations due to geopolitical and
         logistical events, such as Russia’s invasion of Ukraine. These can disrupt global energy systems
         and affect the stability of nations and human livelihoods [5,6]. On the other hand, renewable energy
         production and distribution tend to lie within a country’s boundaries. Over the last few years, the
         price of renewable energy has been catching up with those of subsidised fossil fuels – with some
         specific examples already undercutting fossil fuel prices [7]. In fact, from 1987 to 2015, the cost of
         oil and coal rose by approximately 36% and 81%, respectively, and from 1989 to 2015, the cost
         of natural gas rose by approximately 53% [8]. More recently in March 2022, the UK experienced
         increases in natural gas to around £5.40/therm, a rise above 1,100% from the price levels seen
         in 2021 [9]. Nevertheless, it is important to note that renewable energy variability and investment
         requirements are significant challenges to grid stability and energy security.
         There are good reasons for this decision. First, the IMO focuses mainly on ships that navigate
         international waters or large ships performing domestic voyages [14]. These vessels are required
         to have the automatic identification system (AIS) transponders for safe navigation. On the other
         hand, small boats tend not to have an AIS or a global positioning system (GPS) transponder
         [15,16], which makes the study of their movements more challenging. Second, small boats are
         typically registered and monitored by national and regional bodies, and the comprehensiveness
         of data depends on capital and human resources in addition to the infrastructure to maintain the
         registry [17]. Third, small boats are a diverse segment of shipping and usage depends on the
         geographical location, type of activity, construction and operating costs and accessibility to fuel or
         bunkering infrastructure [18]. Similarly, engine providers are extensive, giving a broad range of fuel
         consumption curves and emissions [19–21].
         Furthermore, fuel selection is equally diverse: petrol, diesel, petrol mixed with engine oil – mainly for
         two-stroke engines, ethanol and bio-fuels – or a mix of bio-fuel with different fossil fuels. Finally, not
2 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
         https://doi.org/10.14324/111.444/ucloe.000058
                                      BoatNet: automated small boat composition detection using deep learning on satellite imagery
         all small boats are powered by an internal combustion engine. They can instead be powered by sail,
         battery-electric or paddles [22–24].
         Nevertheless, with all these challenges, the small boat fleet can significantly contribute to the
         shipping segment’s emission footprint based on its activity [25,26]. Emissions inventories aid our
         understanding of what measures must be taken to enable governments and industry to start the
         road to full decarbonisation in a just and equitable way [27–29]. Furthermore, creating effective
         policies and regulations based on accurate emissions accounting can incentivise the use of
         energy-efficient technologies, electrification and scalable zero-emission fuels [30,31]. Additionally,
         if countries want to meet their ambitious decarbonisation emissions targets, they cannot afford to
         ignore the role played in greenhouse gas (GHG) emissions by the small boat fleet [32–35].
         Although it is possible to estimate emissions from large vessels using AIS data sent from a ship’s
         transponder to be coupled with technical models [11], small vessels depend on the national
         registration system. Their operation is typically assumed or captured by national fuel sales, which
         tend to be highly aggregated (e.g., [36]). Developed economies, such as the UK, tend to have a
         national registry of smaller vessels [37] that provides a sense of their activity level and hence can
         infer CO2 emissions.
         However, in developing countries, it tends to be a mixed bag in terms of the level of precision and
         availability. For instance, in Mexico, only fishing vessels are counted in the national registry [38].
         Still, it is not easy to know where they are located and infer their activities. Overall, Mexico does
         not have a regional CO2 inventory specialised in the small boat fleet; instead, they are aggregated
         as part of the maritime and fluvial navigation [1A3d] class in the national annual emission inventory
         developed by the Instituto Nacional de Ecologá y Cambio Climático (INECC) [39] in a top-down
         approach based on the IPCC Guidelines [40]. Therefore, quantifying and categorising the small
         boat fleet will allow a better precision of where and how the emissions are being emitted and will
         enhance the maritime emission inventories.
         Observing shipping activity in the Gulf of California is essential due to its unique geographical
         location, conformation and biophysical environment [41–43]. Furthermore, the Gulf of California,
         includes the largest fishing state (Sonora) in Mexico [44] and the most prominent sports fishing
         destination (Los Cabos, Baja) [45]. Additionally, the region is one of the most protected areas
         in Mexico due to its diversity of flora and fauna; the area includes the upper part of the Gulf of
         California, Bahia Loreto and Bahia de los Angeles [46,47].
         Deep learning is a type of artificial intelligence (AI) that has shown great promise in solving complex
         problems in fields such as computer vision and natural language processing. It involves training large
         neural networks on vast amounts of data, which allows them to automatically learn complex patterns
         and relationships [54]. CNNs are a type of deep learning model that is particularly well-suited for
         image recognition tasks. They can learn hierarchical representations of visual data and can handle
         large amounts of data, making them efficient and effective for satellite image recognition [55,56].
         Recent advances in satellite image recognition using deep learning have shown promising results.
         For example, researchers have used CNNs to detect objects or patterns in satellite imagery with high
         accuracy, such as roads, buildings and vegetation [57,58]. They have also applied deep learning to
         tasks such as land use classification, land cover mapping and disaster damage assessment [59–61].
         In conclusion, bringing deep learning, especially CNNs, to the field of satellite image recognition
         is a large area of opportunity. It allows leveraging the power of AI to automatically learn complex
         patterns and relationships in satellite imagery. This can lead to improved accuracy, efficiency,
         automation and scalability compared to traditional methods, and has the potential to benefit a
         range of fields that rely on satellite imagery data.
3 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
         https://doi.org/10.14324/111.444/ucloe.000058
                                      BoatNet: automated small boat composition detection using deep learning on satellite imagery
         Contributions
         The contributions of this study are summarised as follows:
         •     A purpose-built methodology for this work, BoatNet, was developed. This work shows that
               BoatNet detects many small boats in low-resolution, blurry satellite images with considerable
               noise levels. As a result, the precision of training can be up to 93.9%, and detecting small
               boats in the Gulf of California can be up to 74.0%.
         •     This work demonstrated that BoatNet could detect the length of small boats with a precision
               up to 99.0%.
         •     BoatNet has allowed for a better understanding of the small boat activity and physical
               characteristics. Based on this, it has been possible to answer questions about the composition
               of small boats in the Gulf of California. Regarding the authors’ knowledge of the literature, this
               is a first but essential step in constructing a way, based in object recognition, to estimate the
               maritime carbon footprint of the small boat fleet.
         Related work
         Small boat fleet and carbon emissions
         Previous work related to estimating small-scale vessels without machine learning methods includes
         using top-down and bottom-up approaches and the use of statistical assumptions.
         Parker et al. [62] used a top-down approach to estimate fishing sector emissions in 2011, which
         reached about 179 Mt carbon dioxide equivalent (CO2e), representing 17.1% of the total large
         fishing ship emissions in that year [63]. However, their work only distinguished between motorised
         and non-motorised fishing vessels. Greer et al. [64] took a bottom-up approach to classify the
         fishing fleet in six different sizes, three below 24 m long. The findings show that the small fishing
         boat fleet in 2016 emitted 47 Mt CO2, about 22.7% of the total fishing fleet. Ferrer et al. [65] used an
         activity-based method using GPS, landing and fuel-used data to estimate the fishing activity around
         the Baja California Peninsula in Mexico. They found that just the small-scale fishing fleet produced
         3.4 Mt of CO2e in 2014. To put this into context, Mexico’s national inventory for the domestic
         shipping sector, but not accounting for fishing activity, in 2014 was recorded at just 2.2 Mt CO2e,
         clearly placing into perspective the role of this fleet segment on national inventories [39].
         Several authors have proposed using AIS to monitor the carbon emissions of the fleet [66–70].
         Johansson et al. [71] proposed a new model Finnish Meteorological Institute - boat emissions and
         activities simulator (FMI-BEAM) to describe leisure boat fleet emissions in the Baltic Sea region
         with over 3000 dock locations, the national small boat registry, AIS data and vessel survey results.
         However, the method cannot cover countries with no national registry for small boats. Besides,
         small boats are not just leisure boats. Ugé et al. [72] estimated global ship emissions with the help
         of data from AIS. They used more than three billion daily AIS data records to create an activity
         database that captured ship size, speed, and meteorological and marine environmental conditions.
         This method is highly dependent on AIS data; however, these transponders are not normally
         installed on board small boats to capture their activity.
         Zhang et al. [73] included unidentified vessels in the AIS-based vessel emission inventory. In
         doing so they developed an AIS-instrumented emissions inventory, including both identified and
         unidentified vessels. In particular, missing vessel parameters for unidentified vessels were estimated
         from a classification regression of similar vessel types and sizes in the AIS database. However, the
         authors did not discuss whether the regression model applies to vessels in most coastal areas.
         Nor did they explore regional vessel diversity in the database, so statistical inferences and levels of
         uncertainty about the applicability of their method to other unidentified vessels in a defined single
         region (e.g., small boats in the Gulf of California, Mexico) cannot be made.
4 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
         https://doi.org/10.14324/111.444/ucloe.000058
                                                                         BoatNet: automated small boat composition detection using deep learning on satellite imagery
                                            In image recognition, a 7 × 7 image, for example, has 49 elements or cells. If ‘X’ is inputted to
                                            the grid, as shown in Fig. 1a, the computer will interpret it as a series of numbers (e.g., zeros
                                            and ones) as seen in Fig. 1b. If each cell is either black or white, for example, black can be
                                            assigned as one while white would be zero, resulting in a 7 × 7 matrix filled with zeros and ones.
                                            After feeding the algorithm as much data as is available, it will be trained to find parameters to
                                            determine if the object is an ‘X’ or not. For example, if it is a grey-scale picture, each number
                                            is neither zero nor one, but rather a grey-scale value from 0 to 255. If it is a colour image, it will
                                            use the red–green–blue (RGB) colour range. Essentially, no matter what the image is, it can be
                                            interpreted as a combination number inside a matrix, this eventually working as the input of the
                                            neural network. The goal of training a neural network is to find the parameters that make the loss
                                            function – it measures how far an estimated value is from its true value – smallest. However, the
                                            method described above is time-consuming and computationally expensive to train real-world
                                            images. Besides, the algorithm will be hard to recognise once the image is dilated, rotated or
                                            changed.
                                            Based on the Neocognitron Model of Fukushima and Miyake [75], LeCun and Bengio [76]
                                            invented a practical method for image recognition, called the convolutional neural network.
                                            The role of convolution is to use a mathematical method to extract critical features from the
                                            image. This is achieved by extracting the features to use a convolution kernel to carry out the
                                            convolution operation. The convolution kernel is a matrix, usually 3 × 3 or 5 × 5. For instance,
                                            if the convolution kernel is 3 × 3, see Fig. 1c, then a convolution operation will be undertaken
                                            with the 7 × 7 ‘X’ matrix (Fig. 1b) and the kernel (Fig. 1c). The operation result is also known as a
                                            feature map (Fig. 1d) [77].
                                            The feature map reinforces the features of the convolution kernel. The 3 × 3 convolution kernel
                                            portrayed in Fig. 1c has only three oblique blocks of pixels that are ones. So if the original 7 × 7
                                            matrix (Fig. 1b) also has diagonal pixel blocks of ones, the number would be extensive when the
                                            convolution operation is complete, which means the desired feature has been extracted. The
                                            smaller the value of the pixel block in the other positions of the feature map (Fig. 1d), the less it
                                            satisfies the feature. In general, different convolution kernels make it possible to achieve different
                                            feature maps.
                                            The next step after convolution is pooling. The pooling method can reduce the feature map size
                                            and maintain similar features to the feature map before the pooling process. Figure 1e shows the
                                            relatively small feature map after pooling the 5 × 5 matrix (Fig. 1d).
                                            The step after pooling is activation. The activation function decides whether the neuron should
                                            be activated by computing the weighted sum and further adding the bias. The essence of the
                                            activation function is to introduce nonlinear factors to solve problems that a linear model cannot
                                            solve [78]. For example, after activating the sigmoid function, each element in the feature map
                                            would be between zero and one, as shown in Fig. 1f.
                                            It is worth noting that the initial convolution kernel may be artificially set. Nevertheless, machine
Figure 1
                                            learning will go backwards to adjust and find the most suitable convolution kernel based on its
From left to right: (a) Letter X in a 7 ×   data. As an image generally has many features, there will be many corresponding convolution
7 image; (b) letter X in a 7 × 7 matrix;    kernels. After many convolutions and poolings, features can be found, including the diagonal
(c) a 3 × 3 convolution kernel; (d) a 5 ×   lines of the image, the contours and the colour features. This information is taken and fed into
5 feature map; (e) a 3 × 3 feature map
after pooling; (f) a 3 × 3 feature map      the fully connected network for training, and it is finally possible to determine what the
after activating with sigmoid function.     image is.
5 / 20                                      BoatNet: automated small boat composition detection using deep learning on satellite imagery         UCL OPEN ENVIRONMENT
                                            https://doi.org/10.14324/111.444/ucloe.000058
                                      BoatNet: automated small boat composition detection using deep learning on satellite imagery
         One of the computer vision’s most fundamental and challenging problems is target detection. The
         main goal of target detection is to determine the location of an object in an image based on a large
         number of predefined classes. Deep learning techniques, which have emerged in recent years, are
         a powerful method for learning features directly from data and have led to significant breakthroughs
         in the field of target detection. Furthermore, with the rise of self-driving cars and face detection, the
         need for fast and accurate object detection is growing.
         In 2012, AlexNet, a deep CNN (DCNN) proposed by Krizhevsky et al. [79], achieved record accuracy
         in image classification at the ImageNet Large-Scale Visual Recognition Challenge (ILSRVC), making
         CNNs the dominant paradigm for image recognition. Next, Girshick et al. [80] introduced Region-
         based Convolutional Neural Networks (R-CNN), the first CNN-based object detection method.
         The R-CNN algorithm represents a two-step approach in which a region proposal is generated
         first, and then a CNN is used for recognition and classification. Compared to the traditional sliding
         convolutional window to determine the possible regions of objects, R-CNN uses a selective
         search to pre-extract some candidate regions that are more likely to object in order to avoid
         computationally costly classification and object searches, which makes it faster and significantly
         less computationally expensive [80,81]. Overall, the R-CNN approach is divided into four steps:
         1.    Generate candidate regions.
3. Feed the extracted features into a support vector machine (SVM) classifier.
         However, R-CNN also has drawbacks: the selective search method is slow in generating positive
         and negative sample candidate regions for the training network, which affects the overall speed
         of the algorithm; R-CNN needs to perform feature extraction once for each generated candidate
         region separately; there are a large number of repeated operations which limits the algorithm
         performance [82].
         Since its inception, R-CNN has undergone several developments and iterations: Fast R-CNN,
         Faster R-CNN and Mask R-CNN [83–85]. The improvement of Fast R-CNN is the design of a
         pooling layer structure for the region of interest (ROI). The pooling stage effectively solves the
         R-CNN operation that crops and scales image regions to the same size, speeding up the algorithm.
         Faster R-CNN replaces the selective search method with the region proposal network (RPN) [84].
         The selection and judgment of candidate frames are handed over to the RPN for processing, and
         candidate regions are subjected to multi-task loss-based classification and localisation processes.
         Several CNN-based object detection frameworks have recently emerged that can run faster, have
         a higher detection accuracy, produce cleaner results and are easier to develop. Compared to the
         Faster R-CNN model, the You Only Look Once (YOLO) model can better detect smaller objects,
         that is, traffic lights at a distance [86], which is important when detecting objects in satellite images.
         Also, the YOLO model has a faster end-to-end run time and detection accuracy than the Faster
         R-CNN [86]. Mask R-CNN upgrades the ROI pooling layer of the Fast R-CNN to an ROI align layer
         and adds a branching FCN layer, the mask layer, to the bounding box recognition for semantic
         mask recognition [85]. Thus, the Mask R-CNN is essentially an instance segmentation algorithm,
         compared to semantic segmentation.1 Instance segmentation is a more fine-grained segmentation
         of similar objects than semantic segmentation.
         However, even traditional CNNs can be very useful for large-scale image recognition. For example,
         Simonyan and Zisserman [87] researched the effect of convolutional network depth on its
         accuracy in large-scale image recognition settings. Their research found that even with small
         (3 × 3) convolution filters, significant accuracy is achieved by pushing the depth from 16 to 19
         weight layers.
6 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
         https://doi.org/10.14324/111.444/ucloe.000058
                                      BoatNet: automated small boat composition detection using deep learning on satellite imagery
         In this research, the YOLO framework was selected. It uses a multi-scale detection method, which
         enables it to detect objects at different scales and to adapt to changes in the size and shape of the
         objects being observed [88]. Besides, YOLO is highly effective in detecting small objects with high
         accuracy and precision [89]. This makes it an ideal choice for detecting small objects in satellite
         imagery contexts, such as small boats in coastal waters. Additionally, YOLO is highly scalable,
         making it suitable for use in large-scale applications [90].
         Finally, this study intends to develop the first stages of BoatNet. This image recognition model aims
         at detecting small boats, especially leisure and fishing boats in any sea area which, in turn and with
         further development, could significantly reduce uncertainty in the estimation of small boat fleet
         emission inventories in countries where access to tracking infrastructure, costly satellite databases
         and labour-intensive methodologies are important barriers.
         •     most cities in the Gulf of California do not have enough open-access satellite data in 2018 and
               2021, while many cities have relatively rich satellite data between 2019 and 2020;
         •     there has been a steady increase in the collection of satellite data in the Gulf of California from
               2018 to 2020;
         •     the open-access and high-quality satellite data from Google Earth Pro is not immediately
               available to the public;
         •     differences in data accessibility are still evident among different cities. For example, Guaymas
               in the state of Sonora has rich satellite images in 2019 and 2020. However, other cities, such
               as La Ventana in the state of Baja California Sur, did not appear on Google Earth Pro between
               2019 and 2020.
         For this reason, continuing with the previous strategy of analysing the satellite data for each city in
         the Gulf of California would lead to a relatively large information bias and thus would not achieve
         an effective object detection model. Therefore, the following three cities with the richest data-
         accessibility in Google Earth Pro were chosen as the target areas for this study: Santa Rosalia,
         Loreto and Guaymas (see Fig. 2 for their geographical locations). The number of times captured by
         Google Earth Pro [91] is shown in Fig. 3 with a database of 583 images with timestamps between
         2019 and 2020 for the three Mexican coastal cities.
         Preprocessing
         Each satellite image used for training was manually pre-labelled with a highly precise label box [92].
         The original dataset contained images larger than 9 MB, which is an efficiency burden for neural
         network training, especially when few objects are detected. For this reason, all images were resized
         from 4800 pixels × 2908 pixels to 416 pixels × 416 pixels, with the file sizes reduced to between
         10 KB and 40 KB [93].
         Each satellite image for targeting or testing can be directly extracted from Google Earth Pro. Before
         downloading these images, a few things were done initally. First, all the layers from Google Earth
         Pro needed to be removed. Then, it was necessary to open the ‘Navigation’ tab of the ‘Preferences’
         menu; ‘Do not automatically tilt while zooming’ needs to be clicked. This allowed the images
7 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
         https://doi.org/10.14324/111.444/ucloe.000058
                                                                        BoatNet: automated small boat composition detection using deep learning on satellite imagery
Figure 2
Figure 3
                                           available to be acquired which were directly above sea level. Finally, the eye altitude was set to
                                           200 m and the images were saved in 4800 pixels × 2908 pixels.
8 / 20                                     BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                           https://doi.org/10.14324/111.444/ucloe.000058
                                                                     BoatNet: automated small boat composition detection using deep learning on satellite imagery
                                        Satellite images often contain noise such as shadows cast by water on the sea surface or haze
Figure 5                                clouds in the atmosphere, which make the training data inaccurate and often cause problems
Average Precision (AP) versus GPU
                                        ensuring the model’s correctness. He et al. [95] proposed a simple but effective image prior-dark
Speed in the 6th generation of YOLOv5   channel before removing haze from a single input image. The prior-dark channel can be used as a
model under COCO data set [86,94].      statistic of outdoor haze-free images. Based on critical observation, most local patches in outdoor
                                        haze-free images contain some pixels whose intensity is very low in at least one colour channel.
                                        Using this prior-dark channel before the haze imaging model, the thickness of the haze can be
                                        estimated, and a high-quality haze-free image can be recovered. Moreover, a high-quality depth
                                        map can also be obtained as a byproduct of haze removal. In the same way, shadows can be
                                        removed using the prior-dark channel.
                                        Similar to the principle of using convolution kernels, specific image kernels can sharpen the
                                        image. While the sharpening kernel does not produce a higher-resolution image, it emphasises the
                                        differences in adjacent pixel values, making the image appear more vivid. Overall, sharpening an
                                        image can significantly improve its recognition accuracy with a 5 × 5 image kernel.
                                        As the dataset for the training model was created with each edge tangent to the edge of the
                                        detected object, it can roughly treat the boat’s length as the length of the diagonal within the
                                        detection box. Second, as the scale is central to the detection of the small boat fleet, the imagery
                                        scale should adhere to the following rules:
• Cannot be too large. The image should contain the full area in which boats may be found.
                                        •     Cannot be too small. If this is not followed it is highly probable that the group of boats are
                                              detected as a single but larger boat.
                                        •     Be sufficiently clear. This characteristic allows the algorithm to quantify the boat’s length and
                                              accurately classify the measurements.
                                        The eye altitude was set to 200 m based on the above rules. This study used a satellite image of
                                        Zurich Lake, Switzerland, on 16 August 2018 as the standard image for defining the scale (Fig. 6).
9 / 20                                  BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                        https://doi.org/10.14324/111.444/ucloe.000058
                                                                       BoatNet: automated small boat composition detection using deep learning on satellite imagery
                                          Compared with other regions, the satellite image of Zurich Lake complies with the rules, and it is a
Figure 6                                  suitable candidate as the standard for measuring the length of small boats. This standard was then
An image from Google Earth Pro for
                                          used for the rest of the imagery database.
Zurich Lake on 16 August 2018 when
eye alt is 200 m. (Source: Google Earth
                                          A length measurement comparison was made between the BoatNet framework and Google Earth
Pro, 2021.)                               Pro to validate the length method. The vessel in Fig. 6 had a YOLO length of 0.43 that, after the
                                          scale conversion, represented 55.17 m. With an eye altitude of 200 m, the ratio of the absolute
                                          length to the YOLO length was approximately 127. Finally, after several verification tests, this ratio
                                          returned a small margin of error (around 1–3%) and hence was deemed a suitable scaling ratio
                                          for the remaining images. Moreover, having the same ratio and eye altitude was not enough. The
                                          resolution of each image must be the same, so measurements are standardised. For this purpose,
                                          all datasets that detect small boats will maintain a resolution of 3840 × 2160 pixels.
                                          In a certain sense, large vessels (e.g., cargo ships) and small vessels (e.g., small boats for domestic
                                          use) are distinguished when creating the dataset for the area of interest. However, due to the
                                          scaling, some large vessels such as general cargo ships do not appear fully in an image. Hence
                                          they are not considered in the statistical results of this work. On the other hand, some of the larger
                                          vessels, slightly shorter in length than the 200 m eagle eye scale, are identified correctly by the
                                          algorithm and counted as part of the number of large vessels in the area. A Python script was then
                                          designed to count the number of small and large boats between regions.
                                          After distinguishing between large and small boats, it is necessary to distinguish between small
                                          recreational boats for domestic use and fishing boats. The model used the detected deck colour of
                                          the small boat to distinguish between them. If the deck was predominantly white, it was assigned
                                          as a recreational boat, while any other mix of colour would be designated as a fishing boat. It is
                                          recognised that this is a broad and simplistic classification method, but it is an effective one to test
                                          the categorisation power of the model. Each of the detected boats (i.e., the objects within the four
                                          coordinate anchor box) were analysed whether the colour was white or mainly white to assign it to
                                          each category. As a final step, the model performed the category counting for each image, where a
                                          Python script was designed to count the small white boats.
                                          Results
                                          Train custom data: weights, biases logging, local logging
                                          As shown in Fig. 7, the average accuracy, precision and recall of the model all show a significant
                                          increase with the model training number when the intersection over union (IOU)3 is between 50%
10 / 20                                   BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                          https://doi.org/10.14324/111.444/ucloe.000058
                                                                  BoatNet: automated small boat composition detection using deep learning on satellite imagery
a b
c d
                                     and 95%. In particular, the precision of the model can eventually reach a level close to 96%.
Figure 7                             However, this does not necessarily mean that the model will also fit the satellite imagery of the
Average model precision when IOU
                                     Gulf of California. First, such high-accuracy results only tell us that the model can achieve a
is larger than 0.50; average model   relatively high recognition accuracy, which gradually increases and reaches 96% after 300 training
precision when IOU is between 0.50   repetitions. In the case that the algorithm needs to be trained for this area, consideration must be
and 0.95; The model precision; The
model recall rate.
                                     given to purposefully selecting many small boats in or near the area as a data source for training the
                                     model.
                                     To train models faster, the images’ resolution was reduced by about 70 times, resulting in images
                                     of 416 pixels × 416 pixels. The training could otherwise take two weeks if the images used had a
                                     resolution of 4800 pixels × 2908 pixels.
                                     Similarly, as shown in Fig. 8, the loss rate of the box can eventually reach 1% as the
                                     number of training sessions increases. As this study has defined only one class of object
                                     (i.e., boat), the probability that the detection box does not detect that it is a boat at all is 1%.
                                     Similarly, because there is only one class, the class loss rate is zero. Figure 9 presents the
                                     prediction results during the training of the model, and shows that the model can detect the
                                     presence of vessels in 100% of the tested ranges and gives the corresponding range boxes.
                                     Most detected boats have a 90% probability of being boats, an acceptable value for object
                                     detection. As only one class was set, some were also considered a 100% probability of being
                                     boats.
11 / 20                              BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                     https://doi.org/10.14324/111.444/ucloe.000058
                                                                        BoatNet: automated small boat composition detection using deep learning on satellite imagery
a b c
                                           measured 6.98 m using Google Earth Pro, while BoatNet estimated 6.74 m. The error between them
Figure 8
                                           is 3.4%. Google Earth Pro measured the larger ship at 41.38 m and BoatNet at 40.98 m. The error
The box loss rate of the model; the        between them is less than 1.0%.
class loss rate of the model; the object
loss rate of the model.                    As explained in the section Single object dectection architecture, it was unproductive for the model
                                           to select the entire region for the study due to the varying amount of publicly available regional
                                           images from Google Earth Pro over the past three years. Ultimately, the satellite image database
                                           was built from 690 images. However, as stated in that section, some of the slightly earlier satellite
                                           images offered inferior detail representation capabilities, which resulted in the model not accurately
                                           detecting the target’s features. To improve model accuracy, an image enhancement process using a
                                           5 × 5 sharpening kernel allowed for a higher recognition rate. However, the following situations still
                                           occur:
                                           1.    Figure 11: When the detailed representation of the image is indigent, that is, the images are
                                                 blurred, and two or three small boats are moored together, the model is very likely to recognise
                                                 the boats as a whole. There are two reasons for this problem. First, the training data is primarily
                                                 a ‘fuzzy’ data source. Thus, when two or three small boats are moored together, the model
                                                 cannot easily detect the features of each small boat individually. In contrast, it may seem more
                                                 reasonable to the model that the boats as a whole have the same features. The second reason
                                                 is that most data sources are individual boats on the surface or boats docked close to each
                                                 other. As the data sources do not fully consider the fuzzy nature of the detail needed to detect
                                                 the object and the fact that they are too close together, the model naturally does not recognise
                                                 such cases.
                                           2.    Figure 12: When a large cargo ship is moored, the ship appears as a ‘rectangle’ from the air,
                                                 much like a long pier, and is sometimes undetectable because small vessels with a rectangular
                                                 shape were not common at the time the data feed was compiled. This also applies to
                                                 uncommon vessels such as battleships. This could be corrected if the model considered larger
                                                 ships, but this was outside the scope of this work.
                                           3.    Figure 13: The recognition rate was also significantly lower when the boats sometimes lay on
                                                 the beach rather than floating on the water. This is because most of the training data are based
                                                 on images in the water rather than boats on the beach.
                                           Nevertheless, as Figs 11–13 demonstrate, the model still detects most small boats in poorly
                                           detailed satellite images, even those that the human eye cannot easily detect. The number of small
                                           and large ships between regions can be seen in Figs 14 and 15. Two different types of port cities
                                           are exemplified by Guaymas, Loreto and Santa Rosalia:
                                           1.    Santa Rosalia and Loreto have a much smaller number of small boats and almost no large
                                                 ships.
                                           2.    The port of Guaymas presented a larger number of small boats when contrasted to the other
                                                 two coastal cities. There were between 1.37 and 8.00 times more small boats detected in
                                                 Guaymas than in Santa Rosalia (dependent on the month and year of the image) and 3.00
                                                 times more than in Loreto.
                                           3.    Guaymas also had a larger number of large ships. There were more than 10 times more large
                                                 ships detected in Guaymas than in Santa Rosalia and Loreto.
                                           4.    In the relatively large seaports of Guaymas and Loreto, there is a tendency for both large and
                                                 small vessels to decrease with time.
12 / 20                                    BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                           https://doi.org/10.14324/111.444/ucloe.000058
                                                                     BoatNet: automated small boat composition detection using deep learning on satellite imagery
                                        In fact, according to statistics [96] from the Mexican government, in 2020, the populations of
Figure 9                                Guaymas, Loreto and Santa Rosalia were 156,863, 18,052 and 14,357, respectively. Therefore,
                                        the results infer that the number of detected boats is correlated to the number of habitats, which
Test result of a trained model for
detecting ships. (Source: Author-
                                        makes sense as, by probability, there would be more economical and leisure activities around larger
originated, based on Google Earth Pro   coastal cities.
imagery, 2021).
13 / 20                                 BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                        https://doi.org/10.14324/111.444/ucloe.000058
                                                                      BoatNet: automated small boat composition detection using deep learning on satellite imagery
Figure 10
                                         Discussion
                                         This study demonstrated the capabilities of a deep learning approach for the automatic detection
Comparing the length of boats using
                                         and identification of small boats in the waters surrounding three cities in the Gulf of California with a
Google Earth ruler and computer vision
algorithm. This example shows the        precision of up to 74.0%. This work used CNNs to identify types of small vessels. Specifically, this
image from Google Earth Pro for Zurich   study presented an image detection model, BoatNet, capable of distinguishing small boats in the
Lake on 16 August 2018 when the
                                         Gulf of California with an accuracy of up to 93.9%, which is an encouraging result considering the
eye alt was 200 m. (Source: Author-
originated, based on Google Earth Pro    high variability of the input images.
imagery, 2021).
                                         Even with the model’s level of performance using large and highly ambiguous training images, it
                                         was found that image sharpening improved model accuracy. This implies that access to better
                                         quality imagery, such as that available through paid-for services, should considerably improve
                                         model precision and training times.
                                         The results of this research have several important implications. First, the study used satellite
                                         data to predict the number and types of ships in three important cities in the Gulf of California.
                                         The resulting analysis can contribute to the region’s shipping fleet composition, level of activity
Figure 11
14 / 20                                  BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                         https://doi.org/10.14324/111.444/ucloe.000058
                                                                         BoatNet: automated small boat composition detection using deep learning on satellite imagery
Figure 12
                                            and ultimately their carbon inventory by adding the emissions produced by the small boat
                                            fleet. Furthermore, through this approach, it is also possible to assign emissions into regions
                                            supporting the development of policies that can mitigate local GHG and air pollution. In addition,
                                            the transfer learning algorithm can be pre-trained in advance and immediately applied to any
                                            sea area worldwide. This will provide a potential method to increase efficiency for scientists
                                            and engineers worldwide who need to estimate local maritime emissions. In addition, the model
                                            can quickly and accurately identify the boat’s length and classify them, allowing researchers to
                                            allocate more time to the vessels they need concentrate on, not just small boats. Finally, all of
                                            the above benefits can be exploited in under served areas with a shortage of infrastructure and
                                            resources.
                                            This work is the first step to building emission inventories through image recognition, and it
                                            has some limitations. The study considered the ship as a single detection object. It did not
                                            evaluate whether the model can improve the accuracy of identifying ships in the case of multiple
Figure 13
15 / 20                                     BoatNet: automated small boat composition detection using deep learning on satellite imagery   UCL OPEN ENVIRONMENT
                                            https://doi.org/10.14324/111.444/ucloe.000058
                                                                        BoatNet: automated small boat composition detection using deep learning on satellite imagery
Figure 14
Figure 15
                                           detection objects. For instance, BoatNet was not trained to detect docks to improve the metrics
                                           of detecting boats. By down-sampling the image to 416 pixels × 416 pixels, it is possible to mask
                                           some of the boats at the edges of the photograph. Furthermore, deep learning models train faster
                                           on small images [97]. A larger input image requires the neural network to learn from four times as
                                           many pixels, increasing the architecture’s training time. In this work, a considerable proportion
                                           of the images in the dataset were large images of 4800 pixels × 2908 pixels. Thus, BoatNet was
                                           set to learn from resized small images measuring 416 pixels × 416 pixels. Due to the low data
                                           quality of the selected regions, the images are less suitable as training datasets. However, using
                                           datasets from other regions or higher-quality open-source imagery may result in inaccurate
                                           coverage of all types of ships in the region. When focusing on the small boat categorisation and
                                           the data used, understanding the implications of different environments (e.g., water or land) on
                                           object classification accuracy through the AI fairness principle deserves further study. From this
                                           point of view, large-scale collection of data sources in the real physical world would be costly and
                                           time-consuming. That said, it is possible that reinforcement learning, or building simulations in
                                           the virtual world, could reduce the negative impact of the environment on object recognition and
                                           thus improve its categorisation precision. Of all these limitations, model detection still achieves
                                           excellent performance in detecting and classifying small boats. To enrich the analysis, one of
                                           Table 1. Detection example. Small boats, large ships and small white boats in Guaymas in 2020
                                                                                       Mar 20                         Jun 20              Sep 20            Dec 20
                                           Small boats                                      323                           283               215                 187
                                           Large ships                                       46                            38                50                  43
                                           Small white boats                                302                           273               210                 174
                                           Shipping boats                                    21                            10                 5                  13
16 / 20                                    BoatNet: automated small boat composition detection using deep learning on satellite imagery       UCL OPEN ENVIRONMENT
                                           https://doi.org/10.14324/111.444/ucloe.000058
                                       BoatNet: automated small boat composition detection using deep learning on satellite imagery
          the future works planned is comparing research results with different algorithms for the same
          problem.
          It is important to remember that BoatNet currently only detects and classifies certain types of small
          boats. Therefore, to estimate fuel consumption and emissions, it is necessary to couple it with small
          boat behaviour datasets [65], typical machinery, fuel characteristics and emission factors unique to
          this maritime segment [39].
          Finally, this work has demonstrated that deep learning models have the potential to identify small
          boats in extreme environments at performance levels that provide practical value. With further
          analysis and small boat data sources, these methods may eventually allow for the rapid assessment
          of shipping carbon inventories.
          Notes
          1 In computer vision, image segmentation is the process of partitioning an image into multiple image segments, also known as
          image regions or image objects.
2 Colab Pro limits RAM to 32 GB while Pro+ limits RAM to 52 GB. Colab Pro and Pro+ limit sessions to 24 hours.
3 IOU is an evaluation metric used to measure the accuracy of an object detector on a particular dataset.
          Acknowledgements
          We thank UCL Energy Institute, University College London for supporting the research, and academic colleagues and fellows
          inside and outside University College London: Andrea Grech La Rosa, UCL Research Computing, Edward Gryspeerdt, Tom
          Lutherborrow.
          References
          [1]   Pörtner H-O, Roberts DC, Tignor M, Poloczanska ES,                   [5]   Bhattacharyya SC. Fossil-fuel dependence and
                Mintenbeck K, Alegra A, et al. Climate change 2022:                        vulnerability of electricity generation: case of
                impacts, adaptation and vulnerability. Cambridge:                          selected European countries. Energy Policy.
                Cambridge University Press; 2022.                                          2009;37(6):2411–20.
          [2]   Raturi AK. Renewables 2019 global status report.                     [6]   DBIS. Russia–Ukraine and UK energy: factsheet.
                Renewable energy policy network for the 21st century.                      Department for Business, Energy & Industrial Strategy,
                2019.                                                                      UK Government. 2022. Available from: https://www.gov.
          [3]   BP. Statistical review of world energy 2021. BP                            uk/government/news/russia-ukraine-and-uk-energy-
                Statistical Review, London, UK; 2021.                                      factsheet.
          [4]   IEA. World energy outlook 2021 [online]. 2021.                       [7]   IRENA. Renewable power generation costs in 2020.
                Available from: https://iea.blob.core.windows.net/                         International Renewable Energy Agency. 2021. Available
                assets/4ed140c1-c3f3-4fd9-acae-789a4e14a23c/                               from: https://www.irena.org/publications/2021/Jun/
                WorldEnergyOutlook2021.pdf.                                                Renewable-Power-Costs-in-2020.
17 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery                UCL OPEN ENVIRONMENT
          https://doi.org/10.14324/111.444/ucloe.000058
                                       BoatNet: automated small boat composition detection using deep learning on satellite imagery
          [8]   BP. Statistical review of world energy 2016. BP                      [26] Halim RA, Kirstein L, Merk O, Martinez LM.
                Statistical Review, London, UK; 2016.                                     Decarbonization pathways for international maritime
                                                                                          transport: a model-based policy impact assessment.
          [9]   Trading Economics. UK natural gas – 2022 data –                           Sustainability. 2018;10(7):2243.
                2020-2021 historical – 2023 forecast – price – quote.
                Trading Economics. 2022. Available from: https://                    [27] Rissman J, Bataille C, Masanet C, Aden N, Morrow
                tradingeconomics.com/commodity/uk-natural-gas.                            WR III, Zhou N, et al. Technologies and policies to
                                                                                          decarbonize global industry: review and assessment
          [10] Walker TR, Adebambo O, Feijoo MCDA, Elhaimer E,                            of mitigation drivers through 2070. Appl Energy.
               Hossain T, Edwards SJ, et al. Environmental effects of                     2020;266:114848.
               marine transportation. In: World seas: an environmental
               evaluation. Elsevier; 2019. p. 505–30.                                [28] Buira D, Tovilla J, Farbes J, Jones R, Haley B, Gastelum
                                                                                          D. A whole-economy deep decarbonization pathway for
          [11] Healy S. Grenhouse gas emissions from shipping: Waiting                    Mexico. Energy Strategy Rev. 2021;33:100578.
               for concrete progress at IMO level. Policy department
               for economic, scientific and quality of life policies                 [29] Adams WM, Jeanrenaud S. Transition to sustainability:
               directorate-general for internal policies. 2020; 19.                       towards a humane and diverse world. International
                                                                                          Union for Conservation of Nature (IUCN); 2008.
          [12] IMO. Adoption of the initial IMO strategy on reduction
               of GHG emissions from ships and existing IMO activity                 [30] Anika OC, Nnabuife SG, Bello A, Okoroafor RE, Kuang
               related to reducing GHG Emissions in the shipping                          B, Villa R. Prospects of low and zero-carbon renewable
               sector. 2018. Available from: https://unfccc.int/sites/                    fuels in 1.5-degree net zero emission actualisation by
               default/files/resource/250_IMO%20submission_                               2050: a critical review. Carbon Capture Sci Technol.
               Talanoa%20Dialogue_April%202018.pdf.                                       2022;5:100072.
          [13] UK Government. Operational standards for small                        [31] Woo J, Fatima R, Kibert CJ, Newman RE, Tian Y,
               vessels. 2021. [Accessed 30 August 2021]. Available                        Srinivasan RS. Applying blockchain technology for
               from: https://www.gov.uk/operational-standards-for-                        building energy performance measurement, reporting,
               small-vessels.                                                             and verification (MRV) and the carbon credit market: a
                                                                                          review of the literature. Build Environ. 2021;205:108199.
          [14] An K. E-navigation services for non-solas ships. Int J
               e-Navigat Maritime Econ. 2016;4:13–22.                                [32] Psaraftis HN, Kontovas CA. CO2 emission statistics
                                                                                          for the world commercial fleet. WMU J Marit Aff.
          [15] Stateczny A. AIS and radar data fusion in maritime                         2009;8(1):1–25.
               navigation. Zeszyty Naukowe/Akademia Morska w
               Szczecinie. 2004. p. 329–36.                                          [33] Lindstad H, Asbjørnslett BE, Strømman AH. The
                                                                                          importance of economies of scale for reductions in
          [16] Vachon P, Thomas S, Cranton J, Edel H, Henschel M.                         greenhouse gas emissions from shipping. Energy Policy.
               Validation of ship detection by the RADARSAT synthetic                     2012;46:386–98.
               aperture radar and the ocean monitoring workstation.
               Can J Remote Sens. 2000;26(3):200–12.                                 [34] Howitt OJ, Revol VG, Smith IJ, Rodger CJ. Carbon
                                                                                          emissions from international cruise ship passengers’
          [17] Gillett R, Tauati MI. Fisheries of the pacific islands:                    travel to and from New Zealand. Energy Policy.
               regional and national information. FAO Fisheries and                       2010;38(5):2552–60.
               Aquaculture Technical Paper. 2018;625:1–400.
                                                                                     [35] Dalsøren SB, Eide M, Endresen Ø, Mjelde A, Gravir G,
          [18] Calderón M, Illing D, Veiga J. Facilities for bunkering                    Isaksen IS. Update on emissions and environmental
               of liquefied natural gas in ports. Transp Res Procedia.                    impacts from the international fleet of ships: the
               2016;14:2431–40.                                                           contribution from major ship types and ports. Atmos
                                                                                          Chem Phys. 2009;9(6):2171–94.
          [19] Devanney J. The impact of the energy efficiency design
               index on very large crude carrier design and CO2                      [36] Sener S. Balance nacional de energa 2017. España:
               emissions. Sh Offshore Struct. 2011;6(4):355–68.                           Dirección general de planeación e información; 2018.
          [20] Dedes EK, Hudson DA, Turnock SR. Assessing the                        [37] UK Ship Register. UK small ship registration. 2021.
               potential of hybrid energy technology to reduce                            [Accessed 30 August 2021]. Available from: https://
               exhaust emissions from global shipping. Energy Policy.                     www.ukshipregister.co.uk/registration/leisure/.
               2012;40:204–18.
                                                                                     [38] Mexico National Aquaculture and Fisheries Commission.
          [21] Marty P, Corrignan P, Gondet A, Chenouard R, Hétet                         Registered vessels. 2021. [Accessed 30 August 2021].
               J-F. Modelling of energy flows and fuel consumption                        Available from: https://www.conapesca.gob.mx/wb/
               on board ships: application to a large modern cruise                       cona/embarcaciones_registradas.
               vessel and comparison with sea monitoring data. In:
               Proceedings of the 11th International Marine Design                   [39] INECC. Inventario nacional de emisiones de gases y
               Conference, Glasgow, UK. 2012. p. 11–4.                                    compuestos de efecto invernadero (INEGYCEI). 2020.
          [22] Zhen X, Wang Y, Liu D. Bio-butanol as a new generation                [40] Change I. 2006 IPCC guidelines for national greenhouse
               of clean alternative fuel for SI (spark ignition) and                      gas inventories. Hayama, Kanagawa, Japan: Institute for
               CI (compression ignition) engines. Renew Energy.                           Global Environmental Strategies; 2006.
               2020;147:2494–521.                                                    [41] Lluch-Cota SE, Aragón-Noriega EA, Arregun-Sánchez
          [23] Dwivedi G, Jain S, Sharma M. Impact analysis of                            F, Aurioles-Gamboa D, Jesús Bautista-Romero J,
               biodiesel on engine performance – a review. Renew                          Brusca RC, et al. The Gulf of California: review of
               Sustain Energy Rev. 2011;15(9):4633–41.                                    ecosystem status and sustainability challenges.
                                                                                          Prog Oceanogr [online]. 2007;73(1):1–26. Available
          [24] Korakianitis T, Namasivayam A, Crookes R. Natural-                         from: https://www.sciencedirect.com/science/article/pii/
               gas fueled spark-ignition (SI) and compression-ignition                    S0079661107000134.
               (CI) engine performance and emissions. Prog Energy
               Combust Sci. 2011;37(1):89–112.                                       [42] Munguia-Vega A, Green AL, Suarez-Castillo AN,
                                                                                          Espinosa-Romero MJ, Aburto-Oropeza O, Cisneros-
          [25] Cariou P, Parola F, Notteboom T. Towards low carbon                        Montemayor AM, et al. Ecological guidelines for
               global supply chains: a multi-trade analysis of CO2                        designing networks of marine reserves in the unique
               emission reductions in container shipping. Int J Prod                      biophysical environment of the Gulf of California. Rev
               Econ. 2019;208:17–28.                                                      Fish Biol Fish. 2018;28(4):749–76.
18 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery                UCL OPEN ENVIRONMENT
          https://doi.org/10.14324/111.444/ucloe.000058
                                       BoatNet: automated small boat composition detection using deep learning on satellite imagery
          [43] Marinone S. Seasonal surface connectivity in the gulf of                     learning-derived land cover and land use information.
               california. Estuar Coast Shelf Sci [online]. 2012;100:                       Remote Sens. 2019;11(10):1174.
               133–41. Available from: https://www.sciencedirect.com/
               science/article/pii/S0272771412000169.                                [60] Helber P, Bischke B, Dengel A, Borth D. EuroSAT: a
                                                                                          novel dataset and deep learning benchmark for land use
          [44] Meltzer L, Chang JO. Export market influence on the                        and land cover classification. IEEE J Sel Top Appl Earth
               development of the Pacific shrimp fishery of Sonora,                       Obs Remote Sens. 2019;12(7):2217–26.
               Mexico. Ocean Coast Manag [online]. 2006;49(3):222–
               235. Available from: https://www.sciencedirect.com/                   [61] Talukdar S, Singha P, Mahato S, Pal S, Liou Y-A,
               science/article/pii/S0964569106000329.                                     Rahman A. Land-use land-cover classification by
                                                                                          machine learning classifiers for satellite observations – a
          [45] Hernández-Trejo V, Germán GP-D, Lluch-Belda D,                             review. Remote Sens. 2020;12(7):1135.
               Beltrán-Morales L. Economic benefits of sport fishing
               in Los Cabos, Mexico: is the relative abundance a                     [62] Parker RW, Blanchard JL, Gardner C, Green BS,
               determinant? J Sustain Tour. 2012;161:165–76.                              Hartmann K, Tyedmers PH, et al. Fuel use and
                                                                                          greenhouse gas emissions of world fisheries. Nat Clim
          [46] CNANP. Atlas Interactivo de las Áreas Naturales                            Change. 2018;8(4):333–7.
               Protegidas de México. [Accessed 26 April 2023].
               Availble from: http://sig.conanp.gob.mx/website/                      [63] Smith TW, Jalkanen JP, Anderson BA, Corbett JJ, Faber
               interactivo/atlas/.                                                        J, Hanayama S, et al. Third IMO GHG study 2014,
                                                                                          International Maritime Organization (IMO), London, UK;
          [47] SMARN. Islas y Áreas Protegidas del Golfo de                               2015. Available from: https://greenvoyage2050.imo.
               California. [Accessed 5 April 2022]. Available from:                       org/wp-content/uploads/2021/01/third-imo-ghg-study-
               https://www.gob.mx/semarnat/articulos/islas-y-areas-                       2014-executive-summary-and-final-report.pdf.
               protegidas-del-golfo-de-california-269050.
                                                                                     [64]   Greer K, Zeller D, Woroniak J, Coulter A, Winchester M,
          [48] Manfreda S, McCabe MF, Miller PE, Lucas R, Pajuelo                           Palomares MD, et al. Global trends in carbon dioxide (CO2)
               Madrigal V, Mallinis G, et al. On the use of unmanned                        emissions from fuel combustion in marine fisheries from
               aerial systems for environmental monitoring. Remote                          1950 to 2016. Marine Policy [online]. 2019;107:103382.
               Sens. 2018;10(4):641.                                                        Available from: https://www.sciencedirect.com/science/
                                                                                            article/pii/S0308597X1730893X.
          [49] Chang N-B, Bai K, Imen S, Chen C-F, Gao W.
               Multisensor satellite image fusion and networking for                 [65] Ferrer EM, Aburto-Oropeza O, Jimenez-Esquivel V,
               all-weather environmental monitoring. IEEE Syst J.                         Cota-Nieto JJ, Mascarenas-Osorio I, Lopez-Sagastegui
               2016;12(2):1341–57.                                                        C. Mexican small-scale fisheries reveal new insights
                                                                                          into low-carbon seafood and ‘climate-friendly’ fisheries
          [50] Willis KS. Remote sensing change detection for
                                                                                          management. Fisheries. 2021;46(6):277–87.
               ecological monitoring in united states protected areas.
               Biol Conserv. 2015;182:233–42.                                        [66] Traut M, Bows A, Walsh C, Wood R. Monitoring shipping
                                                                                          emissions via AIS data? Certainly. In: Low Carbon
          [51] Pourbabaee B, Roshtkhari MJ, Khorasani K. Deep
                                                                                          Shipping Conference 2013, London; 2013.
               convolutional neural networks and learning ECG
               features for screening paroxysmal atrial fibrillation                 [67] Johansson L, Jalkanen J-P, Kukkonen J. A
               patients. IEEE Trans Syst Man Cybern Syst. 2017 Jun                        comprehensive modelling approach for the assessment
               1;48(12):2095–104.                                                         of global shipping emissions. In: International technical
                                                                                          meeting on air pollution modelling and its application.
          [52] Kumar MS, Ganesh D, Turukmane AV, Batta U,
                                                                                          Springer; 2016. p. 367–73.
               Sayyadliyakat KK. Deep convolution neural network
               based solution for detecting plant diseases. J Pharm                  [68] Mabunda A, Astito A, Hamdoune S. Estimating carbon
               Negat Results. 2022;13(1):464–71.                                          dioxide and particulate matter emissions from ships
                                                                                          using automatic identification system data. Int J Comput
          [53] Jozdani SE, Johnson BA, Chen D. Comparing deep
                                                                                          Appl. 2014;88:27–31.
               neural networks, ensemble classifiers, and support
               vector machine algorithms for object-based urban                      [69]   Hensel T, Ugé C, Jahn C. Green shipping: using AIS data to
               land use/land cover classification. Remote Sens.                             assess global emissions. Sustainability Management Forum
               2019;11(14):1713.                                                            | NachhaltigkeitsManagementForum. 2020;28:39–47.
          [54] Shen D, Wu G, Suk H-I. Deep learning in medical image                 [70] Han W, Yang W, Gao S. Real-time identification
               analysis. Annu Rev Biomed Eng. 2017;19:221.                                and tracking of emission from vessels based on
                                                                                          automatic identification system data. In: 2016 13th
          [55] Andrearczyk V, Whelan PF. Using filter banks in
                                                                                          international conference on service systems and service
               convolutional neural networks for texture classification.
                                                                                          management (ICSSSM). 2016. p. 1–6.
               Pattern Recognit Lett. 2016;84:63–69.
                                                                                     [71]   Johansson L, Jalkanen J-P, Fridell E, Maljutenko I, Ytreberg
          [56] Kieffer B, Babaie M, Kalra S, Tizhoosh HR.                                   E, Eriksson M, et al. Modeling of leisure craft emissions. In:
               Convolutional neural networks for histopathology image                       International technical meeting on air pollution modelling
               classification: training vs. using pre-trained networks.                     and its application. Springer; 2018. p. 205–10.
               In: 2017 seventh international conference on image
               processing theory, tools and applications (IPTA), IEEE;               [72] Ugé C, Scheidweiler T, Jahn C. Estimation of worldwide
               2017. p. 1–6.                                                              ship emissions using AIS signals. In: 2020 European
                                                                                          Navigation Conference (ENC). 2020. p. 1–10.
          [57] Sherrah J. Fully convolutional networks for dense
               semantic labelling of high-resolution aerial imagery.                 [73] Zhang Y, Fung J, Chan J, Lau A. The significance of
               arXiv preprint arXiv:1606.02585. 2016.                                     incorporating unidentified vessels into AIS-based ship
                                                                                          emission inventory. Atmos Environ. 2019;203:102–13.
          [58] Kampffmeyer M, Salberg A-B, Jenssen R. Semantic
               segmentation of small objects and modeling of                         [74] McCulloch WS, Pitts W. A logical calculus of the ideas
               uncertainty in urban remote sensing images using                           immanent in nervous activity. Bull Math Biophys.
               deep convolutional neural networks. In: Proceedings                        1943;5(4):115–33.
               of the IEEE conference on computer vision and pattern
               recognition workshops. 2016. p. 1–9.                                  [75] Fukushima K, Miyake S. Neocognitron: a self-organizing
                                                                                          neural network model for a mechanism of visual pattern
          [59] Sheykhmousa M, Kerle N, Kuffer M, Ghaffarian S. Post-                      recognition. In: Competition and cooperation in neural
               disaster recovery assessment with machine                                 nets. Springer; 1982. p. 267–85.
19 / 20   BoatNet: automated small boat composition detection using deep learning on satellite imagery                    UCL OPEN ENVIRONMENT
          https://doi.org/10.14324/111.444/ucloe.000058
                                                                     BoatNet: automated small boat composition detection using deep learning on satellite imagery
                                        [76] LeCun Y, Bengio Y. Convolutional networks for images,                 [87] Simonyan K, Zisserman A. Very deep convolutional
                                             speech, and time series. In: The handbook of brain                         networks for large-scale image recognition. CoRR.
                                             theory and neural networks. 1995. p. 1995. Available                       2015;abs/1409.1556.
                                             from: https://citeseerx.ist.psu.edu/document?
                                             repid=rep1&type=pdf&doi=e26cc4a1c717653f3237                          [88] Li Y, Huang H, Chen Q, Fan Q, Quan H. Research on a
                                             15d751c8dea7461aa105.                                                      product quality monitoring method based on multi scale
                                                                                                                        PP-YOLO. IEEE Access. 2021;9:80373–7.
                                        [77] Bouvrie J. Notes on convolutional neural networks.
                                             2006. Available from: http://web.mit.edu/jvb/www/                     [89] Pham M-T, Courtrai L, Friguet C, Lefèvre S, Baussard
                                             papers/cnn_tutorial.pdf.                                                   A. Yolo-fine: one-stage detector of small objects under
                                                                                                                        various backgrounds in remote sensing images. Remote
                                        [78] Lin G, Shen W. Research on convolutional neural                            Sens. 2020;12(15):2501.
                                             network based on improved relu piecewise activation
                                             function. Procedia Comput Sci. 2018;131:977–984.                      [90] Do J, Ferreira VC, Bobarshad H, Torabzadehkashi M,
                                                                                                                        Rezaei S, Heydarigorji A, et al. Cost-effective, energy-
                                        [79] Krizhevsky A, Sutskever I, Hinton GE. Imagenet                             efficient, and scalable storage computing for large-scale
                                             classification with deep convolutional neural networks.                    AI applications. ACM Trans Storage (TOS). 2020;16(4):
                                             Adv Neural Inf Process Syst. 2012;25:1097–105.                             1–37.
                                        [80] Girshick R, Donahue J, Darrell T, Malik J. Rich feature               [91] Lisle RJ. Google Earth: a new geological resource.
                                             hierarchies for accurate object detection and semantic                     Geology Today. 2006;22(1):29–32.
                                             segmentation. In: Proceedings of the IEEE conference
                                             on computer vision and pattern recognition. 2014. p.                  [92] Lutherborrow T, Agoub A, Kada M. Ship detection in
                                             580–87.                                                                    satellite imagery via convolutional neural networks.
                                                                                                                        2018. Available from: https://www.semanticscholar.org/
                                        [81] Uijlings JR, Van De Sande KE, Gevers T, Smeulders AW.                      paper/SHIP-DETECTION-IN-SATELLITE-IMAGERY-VIA-
                                             Selective search for object recognition. Int J Comput                      NEURAL-Lutherborrow-Agoub/2928b63bd05c4e83a4b9
                                             Vis. 2013;104(2):154–71.                                                   5b47857d010d70bfaac7.
                                        [82] Huang J, Rathod V, Sun C, Zhu M, Korattikara A,                       [93] Nelson J. You might be resizing your images incorrectly.
                                             Fathi A, et al. Speed/accuracy trade-offs for modern                       Jan 2020 [online]. Available from: https://blog.roboflow.
                                             convolutional object detectors. In: Proceedings of                         com/you-might-be-resizing-your-images-incorrectly/.
                                             the IEEE conference on computer vision and pattern
                                             recognition. 2017. p. 7310–1.                                         [94] Jocher G, Chaurasia A, Stoken A, Borovec J,
                                                                                                                        NanoCode012, Kwon Y, et al. Ultralytics/YOLOv5: v6.1 –
                                        [83] Girshick R. Fast R-CNN. In: Proceedings of the IEEE                        TensorRT, TensorFlow Edge TPU and OpenVINO Export
                                             international conference on computer vision. 2015. p.                      and Inference [online]. Feb 2022. Available from: https://
                                             1440–8.                                                                    doi.org/10.5281/zenodo.6222936.
                                        [84] Ren S, He K, Girshick R, Sun J. Faster R-CNN:                         [95] He K, Sun J, Tang X. Single image haze removal using
                                             towards real-time object detection with region proposal                    dark channel prior. In: 2009 IEEE conference on computer
Extra information                            networks. arXiv preprint arXiv:1506.01497, 2015.                           vision and pattern recognition. 2009. p. 1956–63.
                                        [85] He K, Gkioxari G, Dollár P, Girshick R. Mask R-CNN.                   [96] INEGI. Population and housing census 2020
UCL Open: Environment is an open             In: Proceedings of the IEEE international conference on                    in Mexico. [Accessed 11 June 2022]. Available
scholarship publication, all previous        computer vision. 2017. p. 2961–9.                                          from: https://www.inegi.org.mx/app/cpv/2020/
versions and open peer review                                                                                           resultadosrapidos/.
                                        [86] Dwivedi P. Yolov5 compared to faster RCNN. Who
reports can be found online in the           wins? 2020. [Accessed 30 August 2021]. Available from:                [97] Tan M, Le Q. Efficientnetv2: smaller models and faster
UCL Open: Environment Preprint               https://towardsdatascience.com/yolov5-compared-to-                         training. In: International conference on machine
server at ucl.scienceopen.com                faster-rcnn-who-wins-a771cd6c9fb4.                                         learning. PMLR; 2021. p. 10096–106.
20 / 20                                 BoatNet: automated small boat composition detection using deep learning on satellite imagery                 UCL OPEN ENVIRONMENT
                                        https://doi.org/10.14324/111.444/ucloe.000058