Single-Species Leaf Detection against Complex Backgrounds with YOLOv5s

Accurate and rapid localization and identification of tree leaves are of significant importance for urban forest planning and environmental protection. Existing object detection neural networks are complex and often large, which hinders their deployment on mobile devices and compromises their effici...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Forests 2024-06, Vol.15 (6), p.894
Hauptverfasser: Wang, Ziyi, Su, Xiyou, Mao, Shiwei
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue 6
container_start_page 894
container_title Forests
container_volume 15
creator Wang, Ziyi
Su, Xiyou
Mao, Shiwei
description Accurate and rapid localization and identification of tree leaves are of significant importance for urban forest planning and environmental protection. Existing object detection neural networks are complex and often large, which hinders their deployment on mobile devices and compromises their efficiency in detecting plant leaves, especially against complex backgrounds. To address this issue, we collected eight common types of tree leaves against complex urban backgrounds to create a single-species leaf dataset. Each image in this dataset contains only one type of tree but may include multiple leaves. These leaves share similar shapes and textures and resemble various real-world background colors, making them difficult to distinguish and accurately identify, thereby posing challenges to model precision in localization and recognition. We propose a lightweight single-species leaf detection model, SinL-YOLOv5, which is only 15.7 MB. First, we integrated an SE module into the backbone to adaptively adjust the channel weights of feature maps, enhancing the expression of critical features such as the contours and textures of the leaves. Then, we developed an adaptive weighted bi-directional feature pyramid network, SE-BiFPN, utilizing the SE module within the backbone. This approach enhances the information transfer capabilities between the deep semantic features and shallow contour texture features of the network, thereby accelerating detection speed and improving detection accuracy. Finally, to enhance model stability during learning, we introduced an angle cost-based bounding box regression loss function (SIoU), which integrates directional information between ground-truth boxes and predicted boxes. This allows for more effective learning of the positioning and size of leaf edges and enhances the model’s accuracy in detecting leaf locations. We validated the improved model on the single-species leaf dataset. The results showed that compared to YOLOv5s, SinL-YOLOv5 exhibited a notable performance improvement. Specifically, SinL-YOLOv5 achieved an increase of nearly 4.7 percentage points in the mAP@0.5 and processed an additional 20 frames per second. These enhancements significantly enhanced both the accuracy and speed of localization and recognition. With this improved model, we achieved accurate and rapid detection of eight common types of single-species tree leaves against complex urban backgrounds, providing technical support for urban forest surveys, urba
doi_str_mv 10.3390/f15060894
format Article
fullrecord <record><control><sourceid>gale_proqu</sourceid><recordid>TN_cdi_proquest_miscellaneous_3153657887</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><galeid>A799431288</galeid><sourcerecordid>A799431288</sourcerecordid><originalsourceid>FETCH-LOGICAL-c254t-a054c837056a04b44c12aec890549a43daa2525b9e85fcec1007919395b442623</originalsourceid><addsrcrecordid>eNpdkU1PwzAMhiMEEtPYgX9QiQscOvLZNscxGCBN2mFw4BRlmVsy2qYkLR__nqAhhLAPtl49r2XLCJ0SPGVM4suSCJzhQvIDNCJSypRLnB_-6Y_RJIQdjiHyQlI-Qou1basa0nUHxkJIlqDL5Bp6ML11baIrbdvQJ3PXdDV8JFfavFTeDe02JO-2f06eVsvVmwgn6KjUdYDJTx2jx8XNw_wuXa5u7-ezZWqo4H2qseCmYDkWmcZ8w7khVIMpZNSl5myrNRVUbCQUojRgCMa5JJJJEVmaUTZG5_u5nXevA4ReNTYYqGvdghuCYkSwLN5W5BE9-4fu3ODbuJ1iOKeMEipxpKZ7qtI1KNuWrvfaxNxCY41robRRn-VSckZoUUTDxd5gvAvBQ6k6bxvtPxXB6vsL6vcL7AtokXXj</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3072321290</pqid></control><display><type>article</type><title>Single-Species Leaf Detection against Complex Backgrounds with YOLOv5s</title><source>MDPI - Multidisciplinary Digital Publishing Institute</source><source>Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals</source><creator>Wang, Ziyi ; Su, Xiyou ; Mao, Shiwei</creator><creatorcontrib>Wang, Ziyi ; Su, Xiyou ; Mao, Shiwei</creatorcontrib><description>Accurate and rapid localization and identification of tree leaves are of significant importance for urban forest planning and environmental protection. Existing object detection neural networks are complex and often large, which hinders their deployment on mobile devices and compromises their efficiency in detecting plant leaves, especially against complex backgrounds. To address this issue, we collected eight common types of tree leaves against complex urban backgrounds to create a single-species leaf dataset. Each image in this dataset contains only one type of tree but may include multiple leaves. These leaves share similar shapes and textures and resemble various real-world background colors, making them difficult to distinguish and accurately identify, thereby posing challenges to model precision in localization and recognition. We propose a lightweight single-species leaf detection model, SinL-YOLOv5, which is only 15.7 MB. First, we integrated an SE module into the backbone to adaptively adjust the channel weights of feature maps, enhancing the expression of critical features such as the contours and textures of the leaves. Then, we developed an adaptive weighted bi-directional feature pyramid network, SE-BiFPN, utilizing the SE module within the backbone. This approach enhances the information transfer capabilities between the deep semantic features and shallow contour texture features of the network, thereby accelerating detection speed and improving detection accuracy. Finally, to enhance model stability during learning, we introduced an angle cost-based bounding box regression loss function (SIoU), which integrates directional information between ground-truth boxes and predicted boxes. This allows for more effective learning of the positioning and size of leaf edges and enhances the model’s accuracy in detecting leaf locations. We validated the improved model on the single-species leaf dataset. The results showed that compared to YOLOv5s, SinL-YOLOv5 exhibited a notable performance improvement. Specifically, SinL-YOLOv5 achieved an increase of nearly 4.7 percentage points in the mAP@0.5 and processed an additional 20 frames per second. These enhancements significantly enhanced both the accuracy and speed of localization and recognition. With this improved model, we achieved accurate and rapid detection of eight common types of single-species tree leaves against complex urban backgrounds, providing technical support for urban forest surveys, urban forestry planning, and urban environmental conservation.</description><identifier>ISSN: 1999-4907</identifier><identifier>EISSN: 1999-4907</identifier><identifier>DOI: 10.3390/f15060894</identifier><language>eng</language><publisher>Basel: MDPI AG</publisher><subject>Boxes ; data collection ; Datasets ; Environmental protection ; Feature maps ; Flowers &amp; plants ; Forestry ; Forests ; Frames per second ; information exchange ; Information processing ; Information transfer ; Learning ; Leaves ; Localization ; Mobile devices ; Modules ; natural resources conservation ; Neural networks ; Object recognition ; Protection and preservation ; rapid methods ; Species ; texture ; Transfer learning ; Trees ; Urban areas ; urban forestry ; Urban forests</subject><ispartof>Forests, 2024-06, Vol.15 (6), p.894</ispartof><rights>COPYRIGHT 2024 MDPI AG</rights><rights>2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><lds50>peer_reviewed</lds50><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed><cites>FETCH-LOGICAL-c254t-a054c837056a04b44c12aec890549a43daa2525b9e85fcec1007919395b442623</cites></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,776,780,27901,27902</link.rule.ids></links><search><creatorcontrib>Wang, Ziyi</creatorcontrib><creatorcontrib>Su, Xiyou</creatorcontrib><creatorcontrib>Mao, Shiwei</creatorcontrib><title>Single-Species Leaf Detection against Complex Backgrounds with YOLOv5s</title><title>Forests</title><description>Accurate and rapid localization and identification of tree leaves are of significant importance for urban forest planning and environmental protection. Existing object detection neural networks are complex and often large, which hinders their deployment on mobile devices and compromises their efficiency in detecting plant leaves, especially against complex backgrounds. To address this issue, we collected eight common types of tree leaves against complex urban backgrounds to create a single-species leaf dataset. Each image in this dataset contains only one type of tree but may include multiple leaves. These leaves share similar shapes and textures and resemble various real-world background colors, making them difficult to distinguish and accurately identify, thereby posing challenges to model precision in localization and recognition. We propose a lightweight single-species leaf detection model, SinL-YOLOv5, which is only 15.7 MB. First, we integrated an SE module into the backbone to adaptively adjust the channel weights of feature maps, enhancing the expression of critical features such as the contours and textures of the leaves. Then, we developed an adaptive weighted bi-directional feature pyramid network, SE-BiFPN, utilizing the SE module within the backbone. This approach enhances the information transfer capabilities between the deep semantic features and shallow contour texture features of the network, thereby accelerating detection speed and improving detection accuracy. Finally, to enhance model stability during learning, we introduced an angle cost-based bounding box regression loss function (SIoU), which integrates directional information between ground-truth boxes and predicted boxes. This allows for more effective learning of the positioning and size of leaf edges and enhances the model’s accuracy in detecting leaf locations. We validated the improved model on the single-species leaf dataset. The results showed that compared to YOLOv5s, SinL-YOLOv5 exhibited a notable performance improvement. Specifically, SinL-YOLOv5 achieved an increase of nearly 4.7 percentage points in the mAP@0.5 and processed an additional 20 frames per second. These enhancements significantly enhanced both the accuracy and speed of localization and recognition. With this improved model, we achieved accurate and rapid detection of eight common types of single-species tree leaves against complex urban backgrounds, providing technical support for urban forest surveys, urban forestry planning, and urban environmental conservation.</description><subject>Boxes</subject><subject>data collection</subject><subject>Datasets</subject><subject>Environmental protection</subject><subject>Feature maps</subject><subject>Flowers &amp; plants</subject><subject>Forestry</subject><subject>Forests</subject><subject>Frames per second</subject><subject>information exchange</subject><subject>Information processing</subject><subject>Information transfer</subject><subject>Learning</subject><subject>Leaves</subject><subject>Localization</subject><subject>Mobile devices</subject><subject>Modules</subject><subject>natural resources conservation</subject><subject>Neural networks</subject><subject>Object recognition</subject><subject>Protection and preservation</subject><subject>rapid methods</subject><subject>Species</subject><subject>texture</subject><subject>Transfer learning</subject><subject>Trees</subject><subject>Urban areas</subject><subject>urban forestry</subject><subject>Urban forests</subject><issn>1999-4907</issn><issn>1999-4907</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNpdkU1PwzAMhiMEEtPYgX9QiQscOvLZNscxGCBN2mFw4BRlmVsy2qYkLR__nqAhhLAPtl49r2XLCJ0SPGVM4suSCJzhQvIDNCJSypRLnB_-6Y_RJIQdjiHyQlI-Qou1basa0nUHxkJIlqDL5Bp6ML11baIrbdvQJ3PXdDV8JFfavFTeDe02JO-2f06eVsvVmwgn6KjUdYDJTx2jx8XNw_wuXa5u7-ezZWqo4H2qseCmYDkWmcZ8w7khVIMpZNSl5myrNRVUbCQUojRgCMa5JJJJEVmaUTZG5_u5nXevA4ReNTYYqGvdghuCYkSwLN5W5BE9-4fu3ODbuJ1iOKeMEipxpKZ7qtI1KNuWrvfaxNxCY41robRRn-VSckZoUUTDxd5gvAvBQ6k6bxvtPxXB6vsL6vcL7AtokXXj</recordid><startdate>20240601</startdate><enddate>20240601</enddate><creator>Wang, Ziyi</creator><creator>Su, Xiyou</creator><creator>Mao, Shiwei</creator><general>MDPI AG</general><scope>AAYXX</scope><scope>CITATION</scope><scope>3V.</scope><scope>7SN</scope><scope>7SS</scope><scope>7X2</scope><scope>8FE</scope><scope>8FH</scope><scope>8FK</scope><scope>ABUWG</scope><scope>AEUYN</scope><scope>AFKRA</scope><scope>ATCPS</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BHPHI</scope><scope>BKSAR</scope><scope>C1K</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>GNUQQ</scope><scope>HCIFZ</scope><scope>M0K</scope><scope>PATMY</scope><scope>PCBAR</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PYCSY</scope><scope>7S9</scope><scope>L.6</scope></search><sort><creationdate>20240601</creationdate><title>Single-Species Leaf Detection against Complex Backgrounds with YOLOv5s</title><author>Wang, Ziyi ; Su, Xiyou ; Mao, Shiwei</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c254t-a054c837056a04b44c12aec890549a43daa2525b9e85fcec1007919395b442623</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Boxes</topic><topic>data collection</topic><topic>Datasets</topic><topic>Environmental protection</topic><topic>Feature maps</topic><topic>Flowers &amp; plants</topic><topic>Forestry</topic><topic>Forests</topic><topic>Frames per second</topic><topic>information exchange</topic><topic>Information processing</topic><topic>Information transfer</topic><topic>Learning</topic><topic>Leaves</topic><topic>Localization</topic><topic>Mobile devices</topic><topic>Modules</topic><topic>natural resources conservation</topic><topic>Neural networks</topic><topic>Object recognition</topic><topic>Protection and preservation</topic><topic>rapid methods</topic><topic>Species</topic><topic>texture</topic><topic>Transfer learning</topic><topic>Trees</topic><topic>Urban areas</topic><topic>urban forestry</topic><topic>Urban forests</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Wang, Ziyi</creatorcontrib><creatorcontrib>Su, Xiyou</creatorcontrib><creatorcontrib>Mao, Shiwei</creatorcontrib><collection>CrossRef</collection><collection>ProQuest Central (Corporate)</collection><collection>Ecology Abstracts</collection><collection>Entomology Abstracts (Full archive)</collection><collection>Agricultural Science Collection</collection><collection>ProQuest SciTech Collection</collection><collection>ProQuest Natural Science Collection</collection><collection>ProQuest Central (Alumni) (purchase pre-March 2016)</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest One Sustainability</collection><collection>ProQuest Central UK/Ireland</collection><collection>Agricultural &amp; Environmental Science Collection</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Natural Science Collection</collection><collection>Earth, Atmospheric &amp; Aquatic Science Collection</collection><collection>Environmental Sciences and Pollution Management</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>ProQuest Central Student</collection><collection>SciTech Premium Collection</collection><collection>Agricultural Science Database</collection><collection>Environmental Science Database</collection><collection>Earth, Atmospheric &amp; Aquatic Science Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Environmental Science Collection</collection><collection>AGRICOLA</collection><collection>AGRICOLA - Academic</collection><jtitle>Forests</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Wang, Ziyi</au><au>Su, Xiyou</au><au>Mao, Shiwei</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Single-Species Leaf Detection against Complex Backgrounds with YOLOv5s</atitle><jtitle>Forests</jtitle><date>2024-06-01</date><risdate>2024</risdate><volume>15</volume><issue>6</issue><spage>894</spage><pages>894-</pages><issn>1999-4907</issn><eissn>1999-4907</eissn><abstract>Accurate and rapid localization and identification of tree leaves are of significant importance for urban forest planning and environmental protection. Existing object detection neural networks are complex and often large, which hinders their deployment on mobile devices and compromises their efficiency in detecting plant leaves, especially against complex backgrounds. To address this issue, we collected eight common types of tree leaves against complex urban backgrounds to create a single-species leaf dataset. Each image in this dataset contains only one type of tree but may include multiple leaves. These leaves share similar shapes and textures and resemble various real-world background colors, making them difficult to distinguish and accurately identify, thereby posing challenges to model precision in localization and recognition. We propose a lightweight single-species leaf detection model, SinL-YOLOv5, which is only 15.7 MB. First, we integrated an SE module into the backbone to adaptively adjust the channel weights of feature maps, enhancing the expression of critical features such as the contours and textures of the leaves. Then, we developed an adaptive weighted bi-directional feature pyramid network, SE-BiFPN, utilizing the SE module within the backbone. This approach enhances the information transfer capabilities between the deep semantic features and shallow contour texture features of the network, thereby accelerating detection speed and improving detection accuracy. Finally, to enhance model stability during learning, we introduced an angle cost-based bounding box regression loss function (SIoU), which integrates directional information between ground-truth boxes and predicted boxes. This allows for more effective learning of the positioning and size of leaf edges and enhances the model’s accuracy in detecting leaf locations. We validated the improved model on the single-species leaf dataset. The results showed that compared to YOLOv5s, SinL-YOLOv5 exhibited a notable performance improvement. Specifically, SinL-YOLOv5 achieved an increase of nearly 4.7 percentage points in the mAP@0.5 and processed an additional 20 frames per second. These enhancements significantly enhanced both the accuracy and speed of localization and recognition. With this improved model, we achieved accurate and rapid detection of eight common types of single-species tree leaves against complex urban backgrounds, providing technical support for urban forest surveys, urban forestry planning, and urban environmental conservation.</abstract><cop>Basel</cop><pub>MDPI AG</pub><doi>10.3390/f15060894</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 1999-4907
ispartof Forests, 2024-06, Vol.15 (6), p.894
issn 1999-4907
1999-4907
language eng
recordid cdi_proquest_miscellaneous_3153657887
source MDPI - Multidisciplinary Digital Publishing Institute; Elektronische Zeitschriftenbibliothek - Frei zugängliche E-Journals
subjects Boxes
data collection
Datasets
Environmental protection
Feature maps
Flowers & plants
Forestry
Forests
Frames per second
information exchange
Information processing
Information transfer
Learning
Leaves
Localization
Mobile devices
Modules
natural resources conservation
Neural networks
Object recognition
Protection and preservation
rapid methods
Species
texture
Transfer learning
Trees
Urban areas
urban forestry
Urban forests
title Single-Species Leaf Detection against Complex Backgrounds with YOLOv5s
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-02T15%3A57%3A30IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-gale_proqu&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Single-Species%20Leaf%20Detection%20against%20Complex%20Backgrounds%20with%20YOLOv5s&rft.jtitle=Forests&rft.au=Wang,%20Ziyi&rft.date=2024-06-01&rft.volume=15&rft.issue=6&rft.spage=894&rft.pages=894-&rft.issn=1999-4907&rft.eissn=1999-4907&rft_id=info:doi/10.3390/f15060894&rft_dat=%3Cgale_proqu%3EA799431288%3C/gale_proqu%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3072321290&rft_id=info:pmid/&rft_galeid=A799431288&rfr_iscdi=true