中文
Share
Share
中文
ISSN 2096-7217 CN 32-1887/S2
Quick Search
Citation Search
Site Search
Adv Search
Toggle navigation
JIAM
Organizer
Home
About
Journal Introduction
Editorial Board
Open Access Statement
Ethic Statement
Academic Misconduct
Browse
Just Accepted
Current Issue
Archive
Hot Papers
Special Issues
Most Cited
Most Viewed
Most Downloaded
Author Center
Online Submission
Guides for Authors
Review Process
Peer Review
Copyright Transfer Agreement
Charge and Remuneration
Download
Manuscript Review
Expert Review
Edit office
Editor in chief
Included
CNKI
WANFANG DATA
VIP
SUPERSTAR
Subscribe
RSS
Email Alert
Subscription
Contact Us
Contact Us
Advertisement
Article Result
Journals
Publication Years
Keywords
Search within results
(((WANG Jinhai[Author]) AND 1[Journal]) AND year[Order])
AND
OR
NOT
Title
Author
Institution
Keyword
Abstract
PACS
DOI
Please wait a minute...
Choose
Download reference
EndNote
Reference Manager
ProCite
BibTeX
RefWorks
Show/Hide thumbnails
Select
Detection of grapes in orchard environment based on improved YOLO-v4
XIAO Zhangna, LUO Lufeng, CHEN Mingyou, WANG Jinhai, LU Qinghua, LUO Shaoming
Journal of Intelligent Agricultural Mechanization 2023, 4 (2): 35-43. DOI:
10.12398/j.issn.2096-7217.2023.02.004
Abstract
(
328
)
PDF
(42180KB)(
187
)
Knowledge map
Aiming at the difficulty for the grape robot to formulate a collision-free picking strategy based on the visual detection results due to the complex and changeable grape growth scene in the orchard environment, a grape detection method based on improved YOLO-v4 in different occlusion states was proposed. First, according to the state of the grape growth scene in the orchard environment, the grapes were marked into four types: no shading, leaf shading, branch shading, and overlapping shading. Then the YOLO-v4 framework was used as the detection model, in which the attention mechanism model (CBAM) was respectively embedded in the backbone network (CSPDarknet53, YOLO-C-C) and path aggregation network (PANet, YOLO-C-P). The network was enhanced by performing target attention on the CSPDarknet53 and PANet network feature extraction processes. Furthermore, the ability to extract grape features reduced the interference of complex scenes in order to achieve high-precision detection of grapes under different occlusions in orchard environments. Finally, the best grape detection model YOLO-C-P for the orchard obscuration scenario was derived by comparing the recognition accuracy and F1 scores of YOLO-C-C and YOLO-C-P networks. The performance evaluation of the method and comparison with other algorithms showed that the accuracy of the YOLO-C-P model was 91.26%, 92.47%, 92.41%, and 90.65% for the grape detection of no shading, leaf shading, branch shading, and overlapping shading, respectively, with the average F1 score of 91.71%. Compared with the same series of models YOLO-v4, YOLO-X-X, YOLO-v5-X, the F1 score increased by 12.62, 8.65, and 5.31 percentage points, respectively. The average time to recognize an image was 0.13s. The research can quickly and effectively identify grapes of no shade, leaf shade, branch shade, and overlapping shade, and can help robots formulate picking strategies (picking order and path planning) in orchard environments so as to avoid collisions caused by occlusions, and consequently provides grape robots with an auxiliary decision-making method for orchard picking.
Reference
|
Related Articles
|
Metrics