EOSDA custom neural network for deforestation detection
  • EOSDA Forest Monitoring

EOSDA Custom Neural Net: Deforestation Detection

When thinking about industries that are associated with rising levels of CO2 emissions, sectors that are “closer to nature” are the last thing that comes to mind. However, research shows that forestry and agriculture are responsible for more than 5.3 Gt of CO2 emitted annually due to deforestation, which makes these industries the second largest source of anthropogenic carbon emission. What’s more, tree felling in tropical forests also reduces the amount of soil organic carbon and endangers biodiversity   . Regulatory bodies at a global scale have already taken action regarding this issue. For example, the EU will restrict market availability and export of products that are associated with deforestation activities .

To stay compliant with international regulations and contribute to environment conservation, it’s crucial to implement means for accurate deforestation detection.

EOS Data Analytics is dedicated to providing space solutions for Earth’s problems. By leveraging GIS data and custom AI algorithms, we strive to preserve the environment we live in. With our expertise in remote sensing and data analysis, we decided to contribute to deforestation detection efforts in the EU by monitoring forest cover changes, identifying illegal logging activities, and supporting conservation initiatives by delivering accurate and timely information.

Our Science team has developed a deep-learning neural network algorithm based on the methodology described in Mapping Tropical Forest Cover and Deforestation with Planet NICFI Satellite Images and Deep Learning in Mato Grosso State (Brazil) from 2015 to 2021 . In this blog post, we are giving a high-level overview of how it works.

The Process Of Creating The Deforestation Detection AI

The decision-making process within computer vision models, which enables them to differentiate between distinct object classes, is highly complex.

EOS Data Analytics is proud to represent the expertise of 60 data scientists including 25 PhDs. To get started with creating a deforestation detection model, our teams of AI and GIS scientists elicited functional requirements. They concluded that the model should:

  • Recognize and filter out the noise (clouds and other visual distortions)
  • Identify change dynamics of deforestation (true positives, true negatives)
  • Identify falsely identified forest/non-forest classes (false positives, false negatives) and run algorithms to determine and assign true values.

To meet these requirements, we had to carefully evaluate our choices regarding source imagery, architectural design, and innovative solutions to address novel algorithmic challenges.

Imagery

Before diving into the intricacies of algorithms, we needed to obtain high-quality satellite imagery datasets to work with.

For the purpose of this research, we have chosen PlanetScope. This source provides mosaics that represent an average best image from multiple ones taken during the month. As a result, they generally include fewer clouds which makes this imagery fairly consistent for efficient training of classification models. Later on, we also experimented with different kinds of channels, tiles, architectures (UNet & MobileNet), and imagery samples, which allowed us to construct an optimal custom neural net. Below, we demonstrate some of the stages of our successful experiments. While the described routine doesn’t represent the entire complexity of the process, it highlights some key findings that showed to be efficient, transferable, and valuable for our future deforestation detection products and services.

Architecture

To conduct the initial segmentation of the area of interest, we used U-Net architecture. It takes raw images and gives a complete segmentation as the output, leveraging the series of convolutional neural networks within itself to study the image features. This type of architecture uses the encoder-decoder logic, shrinking down the images, finding patterns, and then expanding them back, which is why the learning phase looks like a letter “U” .

In our case, the model received a 4-band image (Fig. 1). We found that it benefits from assessing the visible light spectrum complemented by the near-infrared, which is one of the interesting approaches that we found in a series of experiments. At the output, we have a binary mask, where each pixel corresponds to either 0 (not a forest) or 1 (forest).

As you can see in the images above, clouds and noise are classified as non-forest (Fig. 2). In our next step, we will run an algorithm that filters out the clouds from these initial segmentation results.

On the one hand, we do want clouds to be classified as non-forest. Since these are binary masks, it would’ve been worse if they were classified as forest. On the other hand, we don’t know what is beneath the clouds so we cannot track deforestation. That’s why we trained the model to first recognize clouds only to apply the cloud filter that would get rid of the clouds. After that, we were able to train an additional model to distinguish trees from bushes, roads, bare land, and more.

clouds filtered out from a binary mask
Figure 3: Binary masks after applying Cloud Temporal Filter.

After applying this filter, the output looks much cleaner (Fig. 3). That is because the algorithm assigns a forest value to a pixel only if the forest was detected in at least half of the observations in the specified period. The latter was carefully chosen by our team after a series of additional experiments. Eventually, we found the ideal satellite time series range before and after the specified date. Cloud filter essentially refers to those additional images (when there were supposedly no clouds) to recreate the accurate land cover of the questionable area.

EOSDA Forest Monitoring

Combined powers of satellite tech, data analysis, and satisfying user experience, all in one platform to monitor and manage forest resources.

More info

Confidence Scale

At this point, we’ve got clean forest and non-forest binary masks that we will apply in the model’s final deforestation classification process (Fig. 4-6).

Next, we analyze the semi-annual time series and construct a map of felling and forest cover over that period of time. To do that, we calculate the confidence scale (how confident the model is that this pixel belongs to a certain class – forest/non-forest/felling) according to the following rules:

Deforestation Confidence Scale
Confidence score Status Description
1 Confirmed Same class on all 6 masks, e.g., felling, results in felling confirmed
2 High Confidence (Most Likely) Same class, e.g. forest is observed for almost the entire period but is not a forest for one or two dates, and these dates are not the first and last dates of the period. For example, if a pixel is a forest from January to June, but in May, the model detects it as non-forest due to clouds, then we assign this pixel a forest class with high confidence.
3 Unconfirmed Applies only to felling for the latest date, which cannot be confirmed. A felling is unconfirmed when on date N – 1 we had forest, on date N – no forest, and we have no data for date N + 1 to confirm or deny this felling.
deforestation confidence map
Figure 4: Semi-annual deforestation Confidence Scale January-June 2021.

Output Data:

  • Deforestation map, including three classes: forest, non-forest, and felling
  • Confidence Scale, including three scores: result confirmed, high confidence, not confirmed

Here, forest and non-forest classes can have only two statuses – confirmed and most likely. Similarly, a felling class can also have only two statuses – confirmed and unconfirmed. Thus, we can estimate how the forest cover changed in six months by analyzing one binary mask after another from the present to the past, thereby registering the change dynamics.

EOSDA’s Proprietary Cloud Correction For Residual Uncertainties

Although we’ve gotten rid of clouds before, some of them may still be present on semi-annual felling and forest cover maps. We have developed two additional correction algorithms that apply to semi-annual maps.

The decisions they make depend on the model’s observations:

  • Deforest before + forest after = forest before
  • Non-forest before + forest after = forest before

Just like with the confidence scale that we mentioned above, here we can’t confirm the latest dates because we have no data to verify these results.

However, there’s an interesting approach that we can apply if we go back in time. We can verify older maps with previously verified newer maps. For example, a map for December 2022 verifies a map for June 2022, which in turn verifies an even earlier map for December 2021. By doing that, we can retrospectively assess deforestation dynamics over multiple years.

Results Of Training

In the examples below, we used 11 distinct masks for each date throughout our periods of interest. Such methodological choice aimed to optimize the model’s accuracy in classification while reducing potential errors.

results of deforestation detection by a custom machine learning model
Figure 5: Left to right: Deforestation detection mask, satellite image before felling, and after felling. Northeastern Rondônia, Brazil.

With semi-annual maps of felling and forest cover, we can track deforestation over time and create monitoring maps. Such maps allow us to track deforestation volumes over a period of time and compare deforestation rates, trends, etc. In this example above, we created a two-year monitoring map using four semi-annual masks. We can also create such maps for longer periods of time, provided that we have the required imagery.

deforestation tracking
Video showing deforestation monitoring capabilities in EOSDA Forest Monitoring.

These custom solutions can either be sent in reports directly to stakeholders or be implemented within the web interface of EOSDA Forest Monitoring, which is especially relevant given that the new EUDR (European Union’s Deforestation-free Regulation) comes into force by the end of 2024.

deforestation detection results, Brazil
Figure 6: Final result. Deforestation monitoring map for a two-year period 01.2021 - 12.2022. Color-coded zones show deforestation as per 4 semi-annual periods of time. Northeastern Rondônia, Brazil.

The value of deforestation detection models is especially evident when we zoom out and consider tracking large areas. Take a look at the example below (Fig. 7). In this case, tracking the presence or disappearance of trees seems like a humanly impossible task. Meanwhile, it is achievable and has been automated with our machine-learning model.

deforestation detected in Tasmania
Figure 7: Deforestation detection in Tasmania, Sentinel-2.

Availability Of Custom Deforestation Models In The Future

After experimenting with a variety of satellite imagery types, neural net architectures, and algorithms, we found an optimal combination that allows us to continuously detect tree felling with a high degree of confidence.

From a long-term perspective, it is possible to build both retrospective deforestation maps for certain time periods and monitoring maps to assess the rate and volume of deforestation as time goes by. This model can work as part of user-facing applications.

Forest areas are very large, especially in tropical regions. It is virtually impossible to have a big enough team of GIS specialists to look at satellite images and monitor such areas in an attempt to continuously track deforestation. We created this model to simplify this process, providing opportunities to derive data-driven insights. Upon implementation, users will effortlessly monitor the rate, pace, and patterns of deforestation taking place in their area of interest.

Our novel algorithm presents a new and alternative approach for deforestation monitoring, effectively mitigating erroneous predictions linked to various sources of interference, including cloud cover and imperfections in image quality. The outcomes were corroborated using Sentinel-2 images, affirming the neural network’s accurate detection capabilities. These images encompass spatial resolutions and scene partitions that are different from Planet, therefore enhancing the algorithm’s versatility. The resulting neural network architecture is adapted to the use of Sentinel-2 and EOS SAT-1 imagery.

The resulting quality of any machine learning method is very much dependent on the input data, so further evolution of our model can be achieved by utilizing images that don’t require much filtering and false positives/negatives correction.

Considering these factors, exploring alternative sources like the EOS SAT-1 satellite, developed for EOS Data Analytics, presents an opportunity to enhance the input data for more precise and detailed outcomes, given its impressive 1.5-meter per pixel resolution.

Did you enjoy this article?
Thanks for your feedback!

About the author:

Karolina Koval Senior Science Writer at EOS Data Analytics

Karolina is currently pursuing a BSc at Pennsylvania State University. She excels in communicating the scientific value of EOSDA precision and sustainable solutions in an easy-to-read way. Karolina is a dedicated advocate for personal empowerment striving to represent and uplift Ukrainian women in the global STEM community. She’s a member of AWIS, WIT and other organizations.

Recent articles

EOSDA Provides Data Access For Nebraskan Forest Service
  • Case study

EOSDA Provides Data Access For Nebraskan Forest Service

Kun-Yuan Lee, a Conservation Forester GIS specialist at the Nebraska Forest Service, was able to monitor the extensive forest territory remotely with the help of the EOSDA LandViewer.

Growing Tobacco: From Planting To Harvesting And Curing
  • Crop cultivation

Growing Tobacco: From Planting To Harvesting And Curing

Growing tobacco commercially isn't simple — the crop is picky for climate, soil, and nutrients. But lots of resources, practices, and precision farming tools are available to simplify this process.

Mask, Layer, And Indices: New In EOSDA Crop Monitoring
  • EOSDA Crop Monitoring

Mask, Layer, And Indices: New In EOSDA Crop Monitoring

Explore the new features added to EOSDA Crop Monitoring in Q3 2024: cloud/shadow mask management, Planet data integration, new indices, activity recommendations, and mustard growth stages.