Progress 06/15/23 to 06/14/24
Outputs Target Audience:The main target audiences for our project include researchers, plant breeders, industry stakeholders (if applicable), and growers specializing in row crops such as flaxseed, canola, tomato, and strawberry. We intend to make the design and methodologies openly accessible in the public domain. After thorough testing of the technology in research fields, our goal is to facilitate the transfer of this technology to growers' fields for on-farm demonstrations. The aim is to encourage growers to adopt and benefit from the innovative technology in their crop cultivation practices. ? Changes/Problems:The year 2024 was largely successful for crop production in North Dakota, with a few exceptions. The experimental trials for canola and flax were exceptionally well across all five locations. However, the canola and flax experimental plots in Carrington experienced minor hail damage. ? What opportunities for training and professional development has the project provided?A doctoral candidate within Professor Rahman's research group at NDSU is being funded through this grant. This student actively participated in tasks related to flying drones and operating robots both in field and greenhouse settings. Their responsibilities included establishing planned experiments at four different locations and employing suitable experimental designs. Additionally, they were responsible for overseeing the harvest of experimental plots and conducting subsequent statistical analyses. Still at NDSU, this grant has provided opportunities to train two M.S. students in areas related to UASs operations, UAS image collection and processing. ? At UCLA, a Ph.D. student in Computer Science, two Ph.D. students in Mechanical Engineering, and two undergraduate students in Computer Science are being supported in designing and developing unmanned ground vehicles (UGVs) and machine learning models. These efforts focus on using the collected data to extract phenotypes, such as stem counts and diameters, and to predict yield. These projects have become central to their thesis work, forming the core elements of their research. How have the results been disseminated to communities of interest?Video on practical use of Drone and Robot in NDSU Oilseed Breeding program for high-throughput phenotyping have been made public on YouTube: · https://www.youtube.com/watch?v=48V2Y_Hpqxo&t=12s. We have open-sourced both the codebase and dataset for our project, titled "Agronav: An Autonomous Navigation Framework for Agricultural Robots and Vehicles using Semantic Segmentation and Semantic Line Detection." The dataset was collected using an unmanned ground vehicle operating in flax and canola fields located in Fargo and Carrington. It includes high-resolution images captured by a forward-facing RGB camera, along with pixel-level annotations of common object classes within the scenes. Given the labor-intensive nature of creating annotations for semantic segmentation, we anticipate that this dataset will be a valuable resource for computer vision researchers working in agricultural applications. What do you plan to do during the next reporting period to accomplish the goals?The experiment is scheduled for repetition in 2025 at four locations in North Dakota, following the identical experimental design employed in 2022, 2023, and 2024. Along with the field trials, we will continue to use UASs to collect data from the experimental areas and continue to develop approaches to implement high throughput phenotyping approaches in the flax and canola breeding program to support their breeding efforts. We will complete the analysis of 2021 to 2023 data and prepare a manuscript to publish in a peer-reviewed journal.? The primary objective for the upcoming reporting period is to validate the accuracy of the proposed stem diameter estimation pipeline. The development of the pipeline is nearing completion, with the remaining task being to compare the estimated diameters against manually collected measurements. To ensure the robustness of the pipeline under varying environmental conditions, two validation datasets--one collected in a greenhouse and the other in a field setting--will be used. Upon successful validation, we will prepare and submit a manuscript summarizing the findings for publication in a peer-reviewed journal. Additionally, our team will focus on developing a yield prediction model, leveraging both RGB drone imagery and multispectral data. This model will be trained using data collected during the 2023 and 2024 growing seasons.
Impacts What was accomplished under these goals?
Field experiments were carried out during the 2024 growing season across North Dakota, which include locations in Fargo, Prosper, Casselton, Carrington, and Minot. Within the canola breeding nursery at Prosper and Carrington, each site featured two distinct experimental trials: the "Wide Area Trial", consisting of 108 experimental plots; and the "Early Generation Testing", with 460 experimental plots. The canola breeding trial at Minot had 108 experimental plots. The flax breeding trials were located at Fargo, Casselton, and Carrington, each location included three experimental trials: the "Uniform Regional Nursery", with 108 plots; the "Advanced Yield Trial", with 162 plots; and the "Intermediate Yield Trial", with 294 plots. Unmanned aerial systems (UASs) images were collected over the field trials at different times during the growing season. Flights were carried out with a DJI Matrice 300 (M300), outfitted with MicaSense RedEdge-MX Dual system and a DJI Zenmuse P1 cameras, and DJI Mavic 3 multispectral aircrafts. All flights were carried out at 100 ft above ground level. There were six flights conducted in Carrington, while flights in Casselton and Prosper were limited to three, so far, due to late planting caused by excessive rain during planting season. Imagery data from the 2024 growing season is currently being processed. In addition to UAS data collection during the 2024 growing season, during the past year, Co-PD Flores team has been working on processing data to create datasets to support modeling efforts led by the UCLA team. Furthermore, the Flores' lab have been supporting data collection in the greenhouse environment as well, for which they develop a platform to carry to carry both RGB and depth cameras to collect images and videos from canola plants to support the development of machine learning models to automatically measure the width of canola stems. In 2023 and 2024, the small robot was also employed for phenotyping stem diameter both in the field and greenhouse. The process involved capturing a series of RGBD images of plants using the Stereolabs ZED-2i camera, known for its high-resolution depth-sensing capabilities. To accurately measure the stem diameter of canola, a comprehensive machine learning and image processing pipeline was developed and implemented. This end-to-end pipeline encompasses multiple downstream tasks: (1) keypoint detection - which identifies the structure of the stem, (2) semantic segmentation - which isolates individual stems from the background and other stems, (3) skeletonization and pixel pair identification - which determines the accurate pixels for stem diameter estimation, and (4) triangulation and diameter calculation - which uses the depth to calculate the precise stem diameter across the stem. The accuracy and reliability of this pipeline are being further validated on a larger set of plants in 2024, ensuring its robustness and applicability across different environmental conditions. In 2023, Pearson's correlations between these indices and seed yield of canola at 65 days after planting revealed a significant association (0.65****), particularly with indices such as GNDVI531, NDRE, VARI2, GSAVI, NDRE705, OSAVI, GCI, SR, VDVI, and VEG. In 2022, the correlations between vegetation indices and seed yield of canola were highly significant (0.74****) for NDVI, ENDVI, VEG, GRRI, NGRDI, MGRVI, VDVI, VARI, SAVI, and OSAVI at the 66-day after seeding canola. Among the indices, the OSAVI, VDVI, and VEG showed the higher correlation between seed yield of canola and vegetation indices in both 2022 and 2023. The crop harvest of 2024 is in progress. In a similar effort, RGB images were captured using the DJI P1 camera mounted on a drone during multiple flights across different stages of plant growth. This extensive image dataset, collected from over 400 plots between 2022 and 2023, was instrumental in training a deep neural network for yield forecasting of Canola and Flax. An ablation study has been accomplished of different encoder networks: ResNet + LSTM, ResNet + LSTM w/ Contrastive projection, and 3D ResNet-34. We have achieved 12.2 % accuracy in yield prediction using ResNet + LSTM, 10.3 % accuracy using ResNet + LSTM w/ Contrastive projection, and 7.8 % accuracy using 3D ResNet-34. These evaluations were conducted using the 2022 Canola dataset, providing insights into the relative strengths and weaknesses of each model. A similar analysis is currently underway for Flax, where we aim to assess and refine the models further based on the unique characteristics of the crop. The key contribution of the researchers from computer science (CS) in this research lies in the development of advanced machine learning models and frameworks for predicting Canola yields based on drone imagery. The CS researchers, led by UCLA, specifically focused on the design and enhancement of spatiotemporal architectures capable of handling sparse sequential image data, which are typical in agricultural monitoring but distinct from the continuous sequences usually found in video analysis. The CS researchers introduced a sequential combination of spatial and temporal encoders, including a ResNet encoder followed by an LSTM, which is designed to capture both spatial features of images and the temporal relationships between images captured at different times. A significant innovation was the incorporation of contrastive loss, a technique typically used in self-supervised learning, to improve the quality of the model's embeddings. This loss function helps the model to better understand the relationships between similar and dissimilar samples, even when working with limited labeled data, which is common in the agricultural domain. The model architecture and training framework proposed by the CS researchers demonstrated substantial improvements in prediction accuracy. For example, by integrating contrastive loss, the Mean Absolute Percentage Error (MAPE) on Canola yield predictions improved from 13.3% to 10.3% for the 2022 dataset and from 22.5% to 13.4% for the 2023 dataset. Additionally, the model's ability to handle sparse sequences, where images are temporally related but discontinuous, is an important advancement over existing video analysis models that assume frame-level continuity. Furthermore, the CS researchers conducted an ablation study to explore the effects of different encoders, such as ViT and EfficientNet-b0, as well as temporal models like transformers and temporal convolutional networks (TCN). These experiments provided valuable insights into how different architectural choices impact performance in handling agricultural image data, and the results showed that ResNet and LSTM, combined with contrastive loss, achieved the best balance between model complexity and prediction accuracy. Lastly, the CS researchers' work in applying image augmentation techniques, both online (e.g., vertical flipping, color jittering) and offline (e.g., image concatenation), significantly increased the training dataset size, helping to mitigate overfitting in the models and improve generalization to new data. These innovations in model architecture, contrastive learning, and data augmentation underscore the critical role of the CS researchers in advancing yield prediction methodologies in precision agriculture.
Publications
- Type:
Other
Status:
Accepted
Year Published:
2023
Citation:
Rahman M, Flores P, Hosain* S, Jony* M, Hasan* F, Jawed MK, and Wang W (2023) Accelerating Breeding Efficiency by Applying High-Throughput Phenotyping and Genomic Prediction Methods in Canola. An abstract for 2023 International Annual Meetings organized by ASA-CSSA-SSSA, October 29 to November 01, 2023, at St. Louis, MO,USA.
|
Progress 06/15/22 to 06/14/23
Outputs Target Audience:The main target audiences for our project include researchers, plant breeders, industry stakeholders (if applicable), and growers specializing in row crops such as flaxseed, canola, tomato, and strawberry. We intend to make the design and methodologies openly accessible in the public domain. After thorough testing of the technology in research fields, our goal is to facilitate the transfer of this technology to growers' fields for on-farm demonstrations. The aim is to encourage growers to adopt and benefit from the innovative technology in their crop cultivation practices. ? Changes/Problems:In the year 2023, unforeseen natural disasters had a significant impact on our experimental trials. Specifically, at Carrington, the trials had to be abandoned due to the effects of herbicide residuals in the field. The herbicide residual effect posed challenges that led to the decision to discontinue the experimental trials at this location. Additionally, at Casselton, we observed an uneven germination of crops. This irregular germination pattern influenced by uneven moisture introduced variability in the growth and development of the crops at this particular site. In both cases, it is great to have the UAS imagery to help to support those decisions. What opportunities for training and professional development has the project provided?A doctoral candidate within Professor Rahman's research group is being funded through this grant. This student actively participated in tasks related to flying drones and operating robots both in field and greenhouse settings. Their responsibilities included establishing planned experiments at three different locations and employing suitable experimental designs. Additionally, they were responsible for overseeing the harvest of experimental plots and conducting subsequent statistical analyses. A M.S. student in Dr. Flore's lab is being supported through this grant. He is the person heading the efforts related to UAS image collection and the basic image processing (data extraction for individual plots). In addition, that student has assisted Dr. Rahman's students in greenhouse data collection by constructing a rail platform that allows a camera to smoothly move across tables where the plants are growing, improving the quality of the and video collected, which are fundamental to support the work being carried out at UCLA. A Ph.D. student in Computer Science, two Ph.D. students in Mechanical Engineering and 3 undergraduate students in Computer Science at UCLA are being supported to design and develop UGVs and machine learning models using the collected data to extract phenotypes such as stem counts and diameters and to make prediction on yield. These works have become the core elements of their thesis. How have the results been disseminated to communities of interest?Video on practical use of Drone and Robot in NDSU Oilseed Breeding program for high-throughput phenotyping have been made public on YouTube: https://www.youtube.com/watch?v=48V2Y_Hpqxo&t=12s.? What do you plan to do during the next reporting period to accomplish the goals?The experiment is scheduled for repetition in 2024 at three locations in North Dakota, following the identical experimental design employed in both 2022 and 2023. We will complete the analysis of 2023 data and will prepare a manuscript to publish in a peer-reviewed journal. From a UAS image collection perspective, it would be great if we can identify the sensors and times that we should focus on for data collection. That would help us to become more efficient at collecting data, since we could potentially decrease the number of flights, decrease the number of sensors, which in turn would decrease the amount of data collected and processed. In addition, we are currently exploring options that would further increase efficiencies regarding the process and tools that we currently use to draw individual plots on top of the drone image after each flight. That process is currently semi-automated and we looking to increase its level of automation by training a machine learning model to detect the plots based on the drone imagery. At UCLA,we will continue our development of machine learning models for yield prediction using UAS and UGV images. Our methodology encompasses leveraging pre-trained vision models, such as ResNet18 and Vision Transformer, for image encoding. This approach is critical for extracting meaningful features from aerial imagery captured across different plots. Our data processing strategy is bifurcated into two principal components: spatial and temporal information integration. (1) Spatial Information Integration: To address the challenge of capturing comprehensive spatial details of agricultural plots, our dataset includes images shot from varied angles for each plot. This strategy overcomes the limitation of a single image's inability to cover an entire plot effectively. To assimilate this spatial information, we employ contrastive learning techniques. This involves encoding images using the aforementioned pre-trained models to generate representations. These representations are then projected through a feed-forward network, culminating in a loss module designed to minimize intra-plot representational distances while maximizing inter-plot distances. This process ensures that spatially related images contribute cohesively to the plot's representation. (2) Temporal Information Integration: Agricultural plots exhibit significant changes over time, necessitating the incorporation of temporal data for accurate yield prediction. Our dataset, therefore, includes images of each plot captured at regular intervals, specifically during the months of June, July, and August. To integrate this temporal aspect, we initially explore the use of Recurrent Neural Networks (RNN) and Long Short-Term Memory (LSTM) models, known for their efficacy in handling sequential data. Additionally, we consider employing Transformer models, adapted from the time series analysis domain, to enhance our temporal prediction capabilities. By combining these spatial and temporal information processing strategies, we aim to develop a robust framework for agricultural yield prediction. This framework not only capitalizes on the strengths of pre-trained vision models for initial image encoding but also innovatively integrates contrastive learning for spatial information processing and explores advanced models for temporal analysis. Our goal is to achieve a comprehensive understanding of each plot's characteristics over time, thereby enabling more accurate and reliable yield predictions.
Impacts What was accomplished under these goals?
Field experiments were carried out in 2023 at various locations in North Dakota, including Fargo, Prosper, Casselton, Carrington, and Langdon. Within the canola breeding nursery, each site featured three distinct trials: 1) Wide Area Trial - consisting of 108 experimental plots; 2) Advanced Yield Trial - comprising 243 experimental plots, and 3) Early Generation Testing - consisting of 486 experimental plots. Regarding the flax breeding program, each location included three trials: 1) Uniform Regional Nursery with 108 plots; 2) Advanced Yield Trial with 162 plots; and 3) Intermediate Yield Trial with 294 plots. Due to the residual effect of herbicides at the trial site in Carrington, the canola breeding trial at the location had to be abandoned. In 2023, we expanded the scope of unmanned aerial systems (UAS) flights by increasing the number of locations and flight frequency. Flights were carried out with different UASs and cameras to try to capture images that would benefit the goals of the projects. The main aircraft used for the flights was a DJI Matrice 300 aircraft, to which the following sensors were attached, one at the time, to collect images of the trials: DJI Zenmuse P1, Altum-PT, and RedEdge-X Dual Camera System. To check the suitability of a cheaper solution to collect the same type of data, a DJI Mavic 3 Multispectral was used as well. All flights were carried out at 100 ft above ground level (AGL). There were 8 flights over canola and flax trials across all three locations, covering the entire crops growing season. As in previous years, all the UAS collected data was processed by Dr. Flores' team, since they have the resources and expertise to process the data. In summary, the data from the cameras are transferred to a desktop computer, the images are stitched into orthomosaics using Pix4DMapper software from Pix4D, a shapefile containing the IDs of each experimental plot is created, and in-house developed tools are used in ArcGIS Pro to calculate a variety of vegetation indices and extract statistics related to those indices (mean, median, min, max, standard deviation, and range) for each single plot, which then are saved into an Excel file for further analysis. Currently, we are in the process of correlating vegetation indices with agronomic traits across all locations to identify both the best vegetation indices related to traits of interest and the best sensor to be utilized to collect the data. In 2023, the small robot was also employed for phenotyping stem diameter both in the field and greenhouse. The UCLA teamisdeveloping algorithms to convert image data to digital data for precise stem diameter measurements. Additionally, a large robot (5ft long x 5ft wide x 4ft high) has been constructed for whole plot weed control and phenotyping other agronomic traits. We've successfully developed and tested a machine learning-driven computer vision system that accurately measures the stem diameter of canola plants. This system utilizes sophisticated techniques such as semantic segmentation, keypoint detection, and depth estimation, along with various other image processing methods, to precisely determine stem sizes. Currently, this technology is undergoing thorough testing with extensive datasets to ensure its accuracy and reliability before it is shared with relevant communities and stakeholders. In collaboration with Dr. Flores, we conducted four UAV flights (51, 66, 72, and 77 days after seeding) over 675 CANOLA field plots at Carrington during four different growth stages. The images were stitched into ortho-mosaics using Pix4DMapper from Pix4D, then we used an in-house developed Python script to calculate and extract single plot statistics for 38 vegetation indices (VIs) at Dr. Flores's laboratory at NDSU. Pearson's correlations between these indices and seed yield revealed a significant association (0.74****), particularly with indices such as NDVI, ENDVI, VEG, GRRI, NGRDI, MGRVI, VDVI, VARI, SAVI, and OSAVI at the 66-day after seeding canola. Similarly, the UAV flights were conducted over 555 FLAX field plots at Carrington across five different growth stages. Pearson's correlations between these indices and seed yield revealed a significant association (0.70****) with several vegetation indices. This outcome suggests the potential utilization of UAV multispectral images as a proxy for predicting seed yield.
Publications
- Type:
Conference Papers and Presentations
Status:
Published
Year Published:
2023
Citation:
Rahman M, Flores P, Hosain S, Jony M, Hasan F (2023) Accelerating Breeding Efficiency by Applying High-Throughput Phenotyping and Genomic Prediction Methods in Oilseed crops. An abstract for 16th International Rapeseed Congress, September 24-27, 2023, Sydney.
- Type:
Conference Papers and Presentations
Status:
Accepted
Year Published:
2023
Citation:
Delavarpour N, Mathew J, Aduteye J, Flores P. (2023) A comparative study on deep learning algorithms performance on flax crop boll-counting from crop RGB images. 2023 ASABE Annual International Meeting.
- Type:
Conference Papers and Presentations
Status:
Published
Year Published:
2023
Citation:
Rahman M, Flores P, Hosain S, Jony M, Hasan F, Jawed K, Wang W. (2023) Accelerating Breeding Efficiency by Applying High-Throughput Phenotyping and Genomic Prediction Methods in Canola. ASA, CSSA, SSSA International Annual Meeting, 2023, Sydney, Australia.
|