Y-Net: insect counting and segmentation using deep learning on embedded devices
dc.contributor.author | Kargar, Amin | en |
dc.contributor.author | Zorbas, Dimitrios | en |
dc.contributor.author | Gaffney, Michael | en |
dc.contributor.author | O'Flynn, Brendan | en |
dc.contributor.author | Tedesco, Salvatore | en |
dc.contributor.funder | European Regional Development Fund | en |
dc.contributor.funder | Department of Agriculture, Food and the Marine, Ireland | en |
dc.contributor.funder | Teagasc | en |
dc.contributor.funder | Science Foundation Ireland | en |
dc.date.accessioned | 2025-01-09T16:31:40Z | |
dc.date.available | 2025-01-09T16:31:40Z | |
dc.date.issued | 2024 | en |
dc.description.abstract | Insect pests can pose a serious threat to food production and agriculture in general and can cause substantial crop damage and economic losses. Monitoring insect pest populations is essential to control and mitigate these losses. Traditional monitoring methods are considered by growers and agronomists to be time-costly as well as labour-intensive tasks, which ultimately means that in times of high activity on farms it is a task which often is neglected. This study proposes an automated vision-based insect segmentation and counting approach through the use of deep learning (DL) models developed particularly for embedded systems. An image dataset for our target insect, Halyomorpha halys, was first created using images captured by our IoT-enabled image capture system deployed in a fruit orchard. Then, a Y-Net model inspired by UNet was developed with the capability of insect counting in addition to segmentation. The performance of this model was assessed using a variety of different metrics, and the results demonstrated the feasibility and effectiveness of the model in counting and segmentation of insects using Edge-AI algorithms capable of running on embedded systems. Based on the achieved results, the proposed Y-Net model achieved a Mean Squared Error (MSE) of 1.9 for the insect counting task, an Intersection over Union (IoU) of 84.5% and a Dice Similarity Coefficient (DSC) of 91.5% for the segmentation task, with an inference time of nearly 0.4 seconds on a smartphone. | en |
dc.description.sponsorship | Department of Agriculture, Food and the Marine, Ireland (Grant 2020 Trans National ERANET); Teagasc (Walsh Scholarship fund) | en |
dc.description.status | Peer reviewed | en |
dc.description.version | Accepted Version | en |
dc.format.mimetype | application/pdf | en |
dc.identifier.citation | Kargar, A., Zorbas, D., Gaffney, M., O’Flynn, B. and Tedesco, S. (2024) ‘Y-net: insect counting and segmentation using deep learning on embedded devices’, 2024 IEEE International Instrumentation and Measurement Technology Conference (I2MTC). Glasgow, United Kingdom, 20-23 May 2024, pp. 1–6. https://doi.org/10.1109/I2MTC60896.2024.10561007 | en |
dc.identifier.doi | https://doi.org/10.1109/I2MTC60896.2024.10561007 | en |
dc.identifier.endpage | 6 | en |
dc.identifier.uri | https://hdl.handle.net/10468/16800 | |
dc.language.iso | en | en |
dc.publisher | IEEE | en |
dc.relation.ispartof | 2024 IEEE International Instrumentation and Measurement Technology Conference (I2MTC). Glasgow, United Kingdom, 20-23 May 2024 | en |
dc.relation.project | info:eu-repo/grantAgreement/SFI/SFI Research Centres/12/RC/2289_P2/IE/INSIGHT - Irelands Big Data and Analytics Research Centre/ | en |
dc.relation.project | info:eu-repo/grantAgreement/SFI/SFI Research Centres/13/RC/2077/IE/CONNECT: The Centre for Future Networks & Communications/ | en |
dc.relation.project | info:eu-repo/grantAgreement/SFI/SFI Research Centres Programme::Phase 1/16/RC/3835/IE/VistaMilk Centre/ | en |
dc.relation.project | info:eu-repo/grantAgreement/SFI/SFI Research Centres Programme::Phase 1/16/RC/3918/IE/Confirm Centre for Smart Manufacturing/ | en |
dc.rights | © 2024, IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. | en |
dc.subject | Image segmentation | en |
dc.subject | Object counting | en |
dc.subject | CNN-based architecture | en |
dc.subject | Insect monitoring | en |
dc.subject | Precision agriculture | en |
dc.subject | Deep learning | en |
dc.title | Y-Net: insect counting and segmentation using deep learning on embedded devices | en |
dc.type | Article (peer-reviewed) | en |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Y-Net_Insect_Counting_and_Segmentation_using_Deep_Learning_on_Embedded_Devices.pdf
- Size:
- 1.28 MB
- Format:
- Adobe Portable Document Format
- Description:
- Accepted Version
License bundle
1 - 1 of 1
Loading...
- Name:
- license.txt
- Size:
- 2.71 KB
- Format:
- Item-specific license agreed upon to submission
- Description: