Pepper to fall: a perception method for sweet pepper robotic harvesting | Intelligent Service Robotics Skip to main content
Log in

Pepper to fall: a perception method for sweet pepper robotic harvesting

  • Original Research Paper
  • Published:
Intelligent Service Robotics Aims and scope Submit manuscript

Abstract

In this paper we propose a robotic system for picking peppers in a structured robotic greenhouse environment. A commercially available robotic manipulator is equipped with an RGB-D camera used to detect a correct pose to grasp peppers. The detection algorithm uses the state-of-the-art pretrained CNN architecture. The system was trained using transfer learning on a synthetic dataset made with a 3D modeling software, Blender. Point cloud data are used to detect the pepper’s 6DOF pose through geometric model fitting, which is used to plan the manipulator motion. On top of that, a state machine is derived to control the system workflow. We report the results of a series of experiments conducted to test the precision and the robustness of detection, as well as the success rate of the harvesting procedure.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
¥17,985 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price includes VAT (Japan)

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Availability of data and material

The data presented in this work are publically available on the project web site.

References

  1. Orsag M et al (2021) Specularia. https://sites.google.com/view/specularia-pepper-picking, Accessed: 2021-03-31

  2. Bac CW, van Henten EJ, Hemming J, Edan Y (2014) Harvesting robots for high-value crops: state-of-the-art review and challenges ahead. J Field Robot 31(6):888–911

    Article  Google Scholar 

  3. Barth R, Hemming J, van Henten EJ (2016) Design of an eye-in-hand sensing and servo control framework for harvesting robotics in dense vegetation. Biosys Eng 146:71–84

    Article  Google Scholar 

  4. Fu L, Gao F, Wu J, Li R, Karkee M, Zhang Q (2020) Application of consumer rgb-d cameras for fruit detection and localization in field: a critical review. Comput Electron Agric 177:105687

    Article  Google Scholar 

  5. Lin G, Tang Y, Zou X, Xiong J, Fang Y (2020) Color-, depth-, and shape-based 3d fruit detection. Precision Agric 21(1):1–17

    Article  Google Scholar 

  6. Tu S, Pang J, Liu H, Zhuang N, Chen Y, Zheng C, Wan H, Xue Y (2020) Passion fruit detection and counting based on multiple scale faster r-cnn using rgb-d images. Precision Agric 21(5):1072–1091

    Article  Google Scholar 

  7. Wang Z, Walsh KB, Verma B (2017) On-tree mango fruit size estimation using rgb-d images. Sensors 17(12):2738

    Article  Google Scholar 

  8. Zhang J, Karkee M, Zhang Q, Zhang X, Yaqoob M, Fu L, Wang S (2020) Multi-class object detection using faster r-cnn and estimation of shaking locations for automated shake-and-catch apple harvesting. Comput Electron Agric 173:105384

    Article  Google Scholar 

  9. Nguyen TT, Vandevoorde K, Wouters N, Kayacan E, De Baerdemaeker JG, Saeys W (2016) Detection of red and bicoloured apples on tree with an rgb-d camera. Biosys Eng 146:33–44

    Article  Google Scholar 

  10. Perez RM, Cheein FA, Rosell-Polo JR (2017) Flexible system of multiple rgb-d sensors for measuring and classifying fruits in agri-food industry. Comput Electron Agric 139:231–242

    Article  Google Scholar 

  11. Andujar D, Ribeiro A, Fernández-Quintanilla C, Dorado J (2016) Using depth cameras to extract structural parameters to assess the growth state and yield of cauliflower crops. Comput Electron Agric 122:67–73

    Article  Google Scholar 

  12. Milella A, Marani R, Petitti A, Reina G (2019) In-field high throughput grapevine phenotyping with a consumer-grade depth camera. Comput Electron Agric 156:293–306

    Article  Google Scholar 

  13. Lehnert C, English A, McCool C, Tow AW, Perez T (2017) Autonomous sweet pepper harvesting for protected cropping systems. IEEE Robot Autom Lett 2(2):872–879

    Article  Google Scholar 

  14. Kang H, Chen C (2020) Fruit detection, segmentation and 3d visualisation of environments in apple orchards. Comput Electron Agric 171:105302

    Article  Google Scholar 

  15. Arad B, Kurtser P, Barnea E, Harel B, Edan Y, Ben-Shahar O (2019) Controlled lighting and illumination-independent target detection for real-time cost-efficient applications. the case study of sweet pepper robotic harvesting. Sensors,19(6): 1390

  16. Lehnert C, Sa I, McCool C, Upcroft B, Perez T (2016) Sweet pepper pose detection and grasping for automated crop harvesting. In: 2016 IEEE international conference on robotics and automation (ICRA), pp 2428–2434, IEEE

  17. Sa I, Lehnert C, English A, McCool C, Dayoub F, Upcroft B, Perez T (2017) Peduncle detection of sweet pepper for autonomous crop harvesting-combined color and 3-d information. IEEE Robot Autom Lett 2(2):765–772

    Article  Google Scholar 

  18. Hinterstoisser S, Pauly O, Heibel H, Martina M, Bokeloh M (2019) An annotation saved is an annotation earned: using fully synthetic training for object detection. In: Proceedings of the IEEE/CVF international conference on computer vision workshops

  19. Khan S, Phan B, Salay R, Czarnecki K (2019) Procsy: Procedural synthetic dataset generation towards influence factor studies of semantic segmentation networks. In: CVPR workshops, pp 88–96

  20. Di Cicco M, Potena C, Grisetti G, Pretto A (2017) Automatic model based dataset generation for fast and accurate crop and weeds detection. In: 2017 IEEE/RSJ international conference on intelligent robots and systems (IROS), pp 5188–5195, IEEE

  21. Olatunji J, Redding G, Rowe C, East A (2020) Reconstruction of kiwifruit fruit geometry using a cgan trained on a synthetic dataset. Comput Electron Agric 177:105699

    Article  Google Scholar 

  22. Zhang K, Wu Q, Chen Y (2021) Detecting soybean leaf disease from synthetic image using multi-feature fusion faster r-cnn. Comput Electron Agric 183:106064

    Article  Google Scholar 

  23. Barth R, Isselmuiden J, Hemming J, Van Henten EJ (2018) Data synthesis methods for semantic segmentation in agriculture: a capsicum annuum dataset. Comput Electron Agric 144:284–296

    Article  Google Scholar 

  24. Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) Ssd: Single shot multibox detector. In: European conference on computer vision, pp 21–37, Springer

  25. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510–4520

  26. Maric B, Polic M, Tabak T, Orsag M (2020) Unsupervised optimization approach to in situ calibration of collaborative human-robot interaction tools. In: 2020 IEEE international conference on multisensor fusion and integration for intelligent systems (MFI), pp 255–262, IEEE

  27. Hess R (2010) Blender foundations: the essential guide to learning blender 2.6. Focal Press

  28. AliceVision, Meshroom: A 3D reconstruction software., 2018

Download references

Funding

This work has been supported by Croatian Science Foundation under the project Specularia UIP-2017-05-4042 [1].

Author information

Authors and Affiliations

Authors

Contributions

All authors contributed to the study conception and design. Sensing method conception was performed by Marsela Polic and implementation by Jelena Tabak and Marsela Polic. Motion control was implemented and experiments conducted by Marsela Polic and Jelena Tabak. Analysis was performed by all authors. Most of the first draft of the manuscript was written by Marsela Polic. All authors commented and contributed on this manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Marsela Polic.

Ethics declarations

Conflicts of interests

The authors have no relevant financial or non-financial interests to disclose.

Code availability

The code developed and used in this work is publically available within the Laboratory GitHub repository.

Ethics approval

Not applicable.

Consent to participate

Not applicable.

Consent for publication

Not applicable.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Polic, M., Tabak, J. & Orsag, M. Pepper to fall: a perception method for sweet pepper robotic harvesting. Intel Serv Robotics 15, 193–201 (2022). https://doi.org/10.1007/s11370-021-00401-7

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11370-021-00401-7

Keywords

Navigation