Agricultural robots have the prospect to enable more efficient and sustainable agricultural production of food, feed, and fiber. Perception of crops and weeds is a central component of agricultural robots that aim to monitor fields and assess the plants as well as their growth stage in an automatic manner. Semantic perception mostly relies on deep learning using supervised approaches, which require time and qualified workers to label fairly large amounts of data. In this paper, we look into the problem of reducing the amount of labels without compromising the final segmentation performance. For robots operating in the field, pre-training networks in a supervised way is already a popular method to reduce the number of required labeled images. We investigate the possibility of pre-training in a self-supervised fashion using data from the target domain. To better exploit this data, we propose a set of domain-specific augmentation strategies. We evaluate our pre-training on semantic segmentation and leaf instance segmentation, two important tasks in our domain. The experimental results suggest that pre-training with domain-specific data paired with our data augmentation strategy leads to superior performance compared to commonly used pre-trainings. Furthermore, the pre-trained networks obtain similar performance to the fully supervised with less labeled data.
On Domain-Specific Pre- Training for Effective Semantic Perception in Agricultural Robotics / Roggiolani, Gianmarco; Magistri, Federico; Guadagnino, Tiziano; Weyler, Jan; Grisetti, Giorgio; Stachniss, Cyrill; Behley, Jens. - (2023), pp. 11786-11793. (Intervento presentato al convegno IEEE International Conference on Robotics and Automation (ICRA) tenutosi a London; United Kingdom) [10.1109/icra48891.2023.10160624].
On Domain-Specific Pre- Training for Effective Semantic Perception in Agricultural Robotics
Roggiolani, Gianmarco
Software
;Magistri, Federico
Software
;Guadagnino, Tiziano
Conceptualization
;Grisetti, Giorgio
Supervision
;
2023
Abstract
Agricultural robots have the prospect to enable more efficient and sustainable agricultural production of food, feed, and fiber. Perception of crops and weeds is a central component of agricultural robots that aim to monitor fields and assess the plants as well as their growth stage in an automatic manner. Semantic perception mostly relies on deep learning using supervised approaches, which require time and qualified workers to label fairly large amounts of data. In this paper, we look into the problem of reducing the amount of labels without compromising the final segmentation performance. For robots operating in the field, pre-training networks in a supervised way is already a popular method to reduce the number of required labeled images. We investigate the possibility of pre-training in a self-supervised fashion using data from the target domain. To better exploit this data, we propose a set of domain-specific augmentation strategies. We evaluate our pre-training on semantic segmentation and leaf instance segmentation, two important tasks in our domain. The experimental results suggest that pre-training with domain-specific data paired with our data augmentation strategy leads to superior performance compared to commonly used pre-trainings. Furthermore, the pre-trained networks obtain similar performance to the fully supervised with less labeled data.File | Dimensione | Formato | |
---|---|---|---|
Roggiolani_postprint_Domain-Specific_2023.pdf
accesso aperto
Note: https://arxiv.org/pdf/2303.12499 - DOI: 10.1109/ICRA48891.2023.10160624
Tipologia:
Documento in Post-print (versione successiva alla peer review e accettata per la pubblicazione)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
2.56 MB
Formato
Adobe PDF
|
2.56 MB | Adobe PDF | |
Roggiolani_Domain-Specific_2023.pdf
solo gestori archivio
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati (All rights reserved)
Dimensione
2.58 MB
Formato
Adobe PDF
|
2.58 MB | Adobe PDF | Contatta l'autore |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.