no code implementations • 1 Mar 2024 • Katherine Margaret Frances James, Karoline Heiwolt, Daniel James Sargent, Grzegorz Cielniak
Automated phenotyping of plants for breeding and plant studies promises to provide quantitative metrics on plant traits at a previously unattainable observation frequency.
no code implementations • 17 Oct 2023 • Justin Le Louëdec, Grzegorz Cielniak
Selective robotic harvesting is a promising technological solution to address labour shortages which are affecting modern agriculture in many parts of the world.
no code implementations • 21 Sep 2023 • Rajitha de Silva, Grzegorz Cielniak, Junfeng Gao
Vision-based mobile robot navigation systems in arable fields are mostly limited to in-row navigation.
no code implementations • 9 Jun 2023 • Rajitha de Silva, Grzegorz Cielniak, Junfeng Gao
Usage of purely vision based solutions for row switching is not well explored in existing vision based crop row navigation frameworks.
no code implementations • 9 Jan 2023 • Ibrahim Hroob, Sergi Molina, Riccardo Polvara, Grzegorz Cielniak, Marc Hanheide
In this research, we present an end-to-end data-driven pipeline for determining the long-term stability status of objects within a given environment, specifically distinguishing between static and dynamic objects.
no code implementations • 25 Nov 2022 • Riccardo Polvara, Sergi Molina Mellado, Ibrahim Hroob, Grzegorz Cielniak, Marc Hanheide
Long-term autonomy is one of the most demanded capabilities looked into a robot.
no code implementations • 28 Sep 2022 • Rajitha de Silva, Grzegorz Cielniak, Junfeng Gao
We also present a novel crop row detection algorithm for visual servoing in crop row fields.
no code implementations • 23 Sep 2022 • Karoline Heiwolt, Cengiz Öztireli, Grzegorz Cielniak
We present a landmark-free shape compression algorithm, which allows for the extraction of 3D shape features of leaves, characterises leaf shape and curvature efficiently in few parameters, and makes the association of individual leaves in feature space possible.
1 code implementation • 9 Sep 2022 • Rajitha de Silva, Grzegorz Cielniak, Gang Wang, Junfeng Gao
The novel crop row detection algorithm was tested for crop row detection performance and the capability of visual servoing along a crop row.
no code implementations • 4 Apr 2022 • Rajitha de Silva, Grzegorz Cielniak, Junfeng Gao
Our method could reach the performance of a deep learning based crop row detection model trained with real-world data by using 60% less labelled real-world data.
1 code implementation • 7 Dec 2021 • Justin Le Louëdec, Grzegorz Cielniak
In this paper, we propose to employ a Gaussian map representation to estimate precise location and count of 3D surface features, addressing the limitations of state-of-the-art methods based on density estimation which struggle in presence of local disturbances.
1 code implementation • 26 Nov 2021 • Justin Le Louëdec, Grzegorz Cielniak
In this paper, we evaluate modern sensing technologies including stereo and time-of-flight cameras for 3D perception of shape in agriculture and study their usability for segmenting out soft fruit from background based on their shape.
1 code implementation • 21 Sep 2021 • Taeyeong Choi, Owen Would, Adrian Salazar-Gomez, Grzegorz Cielniak
Data augmentation can be a simple yet powerful tool for autonomous robots to fully utilise available data for selfsupervised identification of atypical scenes or objects.
no code implementations • 16 Sep 2021 • Rajitha de Silva, Grzegorz Cielniak, Junfeng Gao
This paper presents a novel metric to evaluate the robustness of deep learning based semantic segmentation approaches for crop row detection under different field conditions encountered by a field robot.
no code implementations • 14 Aug 2021 • Taeyeong Choi, Grzegorz Cielniak
In our previous work, we designed a systematic policy to prioritize sampling locations to lead significant accuracy improvement in spatial interpolation by using the prediction uncertainty of Gaussian Process Regression (GPR) as "attraction force" to deployed robots in path planning.