| Literature DB >> 35075356 |
Amr Abozeid1, Rayan Alanazi1, Ahmed Elhadad1, Ahmed I Taloba1, Rasha M Abd El-Aziz1.
Abstract
Since the Pre-Roman era, olive trees have a significant economic and cultural value. In 2019, the Al-Jouf region, in the north of the Kingdom of Saudi Arabia, gained a global presence by entering the Guinness World Records, with the largest number of olive trees in the world. Olive tree detecting and counting from a given satellite image are a significant and difficult computer vision problem. Because olive farms are spread out over a large area, manually counting the trees is impossible. Moreover, accurate automatic detection and counting of olive trees in satellite images have many challenges such as scale variations, weather changes, perspective distortions, and orientation changes. Another problem is the lack of a standard database of olive trees available for deep learning applications. To address these problems, we first build a large-scale olive dataset dedicated to deep learning research and applications. The dataset consists of 230 RGB images collected over the territory of Al-Jouf, KSA. We then propose an efficient deep learning model (SwinTUnet) for detecting and counting olive trees from satellite imagery. The proposed SwinTUnet is a Unet-like network which consists of an encoder, a decoder, and skip connections. Swin Transformer block is the fundamental unit of SwinTUnet to learn local and global semantic information. The results of an experimental study on the proposed dataset show that the SwinTUnet model outperforms the related studies in terms of overall detection with a 0.94% estimation error.Entities:
Mesh:
Year: 2022 PMID: 35075356 PMCID: PMC8783740 DOI: 10.1155/2022/1549842
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1The architecture of the SwinTUnet model includes encoders, decoders, and skip connections. The Swin Transformer is the base of building the model.
Figure 2Two sequential Swin Transformer blocks.
Figure 3Sample olive images (the first column shows original images, the second column shows ground truth, while the third column shows corresponding detection results).
Overall evaluation of the proposed model.
| No. of images | Total trees | Detected trees | EE (%) | CER | OER |
|---|---|---|---|---|---|
| 230 | 73286 | 72598 | 0.94 | 709 (0.97%) | 912 (1.2%) |
A comparison of the proposed model's results to previous research.
| Technique | Dataset | Spectrum | No. of images | Performance metrics | |||
|---|---|---|---|---|---|---|---|
| OA | CER (%) | OER (%) | EE (%) | ||||
| Reticular matching [ | QuickBird | Greyscale | N/A | 98% | 5 | 7 | 1.24 |
| Multilevel thresholding [ | SIGPAC viewer | Greyscale | 95 | 96% | 5 | 3 | 1.2 |
| Detection using red bands [ | SIGPAC viewer | Red band | 60 | N/A | 1.2 | 4 | 1.27 |
| Improved | SIGPAC viewer | RGB | 110 | 97.5% | 4 | 1 | 0.97 |
| Proposed model | Satellites Pro | RGB | 230 | 98.3% | 0.97 | 1.2 | 0.94 |
Figure 4Performance comparison of the proposed model and the existing work.