Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
13
result(s) for
"Nagasubramanian, Koushik"
Sort by:
Plant disease identification using explainable 3D deep learning on hyperspectral images
by
Singh, Asheesh K.
,
Nagasubramanian, Koushik
,
Ganapathysubramanian, Baskar
in
Agricultural production
,
Agricultural research
,
Artificial intelligence
2019
Background
Hyperspectral imaging is emerging as a promising approach for plant disease identification. The large and possibly redundant information contained in hyperspectral data cubes makes deep learning based identification of plant diseases a natural fit. Here, we deploy a novel 3D deep convolutional neural network (DCNN) that directly assimilates the hyperspectral data. Furthermore, we interrogate the learnt model to produce physiologically meaningful explanations. We focus on an economically important disease, charcoal rot, which is a soil borne fungal disease that affects the yield of soybean crops worldwide.
Results
Based on hyperspectral imaging of inoculated and mock-inoculated stem images, our 3D DCNN has a classification accuracy of 95.73% and an infected class F1 score of 0.87. Using the concept of a saliency map, we visualize the most sensitive pixel locations, and show that the spatial regions with visible disease symptoms are overwhelmingly chosen by the model for classification. We also find that the most sensitive wavelengths used by the model for classification are in the near infrared region (NIR), which is also the commonly used spectral range for determining the vegetative health of a plant.
Conclusion
The use of an explainable deep learning model not only provides high accuracy, but also provides physiological insight into model predictions, thus generating confidence in model predictions. These explained predictions lend themselves for eventual use in precision agriculture and research application using automated phenotyping platforms.
Journal Article
Hyperspectral band selection using genetic algorithm and support vector machines for early identification of charcoal rot disease in soybean stems
by
Singh, Asheesh K.
,
Nagasubramanian, Koushik
,
Ganapathysubramanian, Baskar
in
Agricultural production
,
Agronomic crops
,
Agronomy
2018
Background
Charcoal rot is a fungal disease that thrives in warm dry conditions and affects the yield of soybeans and other important agronomic crops worldwide. There is a need for robust, automatic and consistent early detection and quantification of disease symptoms which are important in breeding programs for the development of improved cultivars and in crop production for the implementation of disease control measures for yield protection. Current methods of plant disease phenotyping are predominantly visual and hence are slow and prone to human error and variation. There has been increasing interest in hyperspectral imaging applications for early detection of disease symptoms. However, the high dimensionality of hyperspectral data makes it very important to have an efficient analysis pipeline in place for the identification of disease so that effective crop management decisions can be made. The focus of this work is to determine the minimal number of most effective hyperspectral wavebands that can distinguish between healthy and diseased soybean stem specimens early on in the growing season for proper management of the disease. 111 hyperspectral data cubes representing healthy and infected stems were captured at 3, 6, 9, 12, and 15 days after inoculation. We utilized inoculated and control specimens from 4 different genotypes. Each hyperspectral image was captured at 240 different wavelengths in the range of 383–1032 nm. We formulated the identification of best waveband combination from 240 wavebands as an optimization problem. We used a combination of genetic algorithm as an optimizer and support vector machines as a classifier for the identification of maximally-effective waveband combination.
Results
A binary classification between healthy and infected soybean stem samples using the selected six waveband combination (475.56, 548.91, 652.14, 516.31, 720.05, 915.64 nm) obtained a classification accuracy of 97% for the infected class. Furthermore, we achieved a classification accuracy of 90.91% for test samples from 3 days after inoculation using the selected six waveband combination.
Conclusions
The results demonstrated that these carefully-chosen wavebands are more informative than RGB images alone and enable early identification of charcoal rot infection in soybean. The selected wavebands could be used in a multispectral camera for remote identification of charcoal rot infection in soybean.
Journal Article
Plant phenotyping with limited annotation: Doing more with less
by
Nagasubramanian, Koushik
,
Ganapathysubramanian, Baskar
,
Singh, Arti
in
Active learning
,
Annotations
,
Classification
2022
Deep learning (DL) methods have transformed the way we extract plant traits—both under laboratory as well as field conditions. Evidence suggests that “well‐trained” DL models can significantly simplify and accelerate trait extraction as well as expand the suite of extractable traits. Training a DL model typically requires the availability of copious amounts of annotated data; however, creating large‐scale annotated dataset requires nontrivial efforts, time, and resources. This limitation has become a major bottleneck in deploying DL tools in practice. Self‐supervised learning (SSL) methods give exciting solution to this problem, as these methods use unlabeled data to produce pretrained models for subsequent fine‐tuning on labeled data and have demonstrated superior transfer learning performance on down‐stream classification tasks. We investigated the application of SSL methods for plant stress classification using few labels. We select a plant stress classification problem to test the effectiveness of SSL, as it is a fundamentally challenging problem due to (a) disease classification which depends on the abnormalities in a small number of pixels, (b) high data imbalance across different classes, and (c) fewer annotated and available plant stress images than in other domains. We compared seven SSL approaches spanning four broad classes of SSL methods on soybean [Glycine max L. (Merr.)] plant stress dataset and report that pretraining on unlabeled plant stress images significantly outperforms transfer learning methods using random initialization for plant stress classification. In summary, SSL‐based model initialization and data curation improves annotation efficiency for plant stress classification tasks and will circumvent data annotation challenges associated with DL methods. Core Ideas Self‐supervised learning (SSL)‐based pretraining provides excellent model initializations. Self‐supervised representations are annotation efficient and transferable for soybean stress classification. Barlow Twins was the best SSL method for annotation efficiency.
Journal Article
Self-supervised maize kernel classification and segmentation for embryo identification
by
Jubery, Talukder Z.
,
Ganapathysubramanian, Baskar
,
Frei, Ursula K.
in
Annotations
,
Classification
,
Computer vision
2023
Computer vision and deep learning (DL) techniques have succeeded in a wide range of diverse fields. Recently, these techniques have been successfully deployed in plant science applications to address food security, productivity, and environmental sustainability problems for a growing global population. However, training these DL models often necessitates the large-scale manual annotation of data which frequently becomes a tedious and time-and-resource- intensive process. Recent advances in self-supervised learning (SSL) methods have proven instrumental in overcoming these obstacles, using purely unlabeled datasets to pre-train DL models.
Here, we implement the popular self-supervised contrastive learning methods of NNCLR Nearest neighbor Contrastive Learning of visual Representations) and SimCLR (Simple framework for Contrastive Learning of visual Representations) for the classification of spatial orientation and segmentation of embryos of maize kernels. Maize kernels are imaged using a commercial high-throughput imaging system. This image data is often used in multiple downstream applications across both production and breeding applications, for instance, sorting for oil content based on segmenting and quantifying the scutellum's size and for classifying haploid and diploid kernels.
We show that in both classification and segmentation problems, SSL techniques outperform their purely supervised transfer learning-based counterparts and are significantly more annotation efficient. Additionally, we show that a single SSL pre-trained model can be efficiently finetuned for both classification and segmentation, indicating good transferability across multiple downstream applications. Segmentation models with SSL-pretrained backbones produce DICE similarity coefficients of 0.81, higher than the 0.78 and 0.73 of those with ImageNet-pretrained and randomly initialized backbones, respectively. We observe that finetuning classification and segmentation models on as little as 1% annotation produces competitive results. These results show SSL provides a meaningful step forward in data efficiency with agricultural deep learning and computer vision.
Journal Article
Self‐supervised learning improves classification of agriculturally important insect pests in plants
by
Abel, Craig A.
,
Singh, Asheesh K.
,
Ganapathysubramanian, Baskar
in
artificial intelligence
,
Classification
,
Datasets
2023
Insect pests cause significant damage to food production, so early detection and efficient mitigation strategies are crucial. There is a continual shift toward machine learning (ML)‐based approaches for automating agricultural pest detection. Although supervised learning has achieved remarkable progress in this regard, it is impeded by the need for significant expert involvement in labeling the data used for model training. This makes real‐world applications tedious and oftentimes infeasible. Recently, self‐supervised learning (SSL) approaches have provided a viable alternative to training ML models with minimal annotations. Here, we present an SSL approach to classify 22 insect pests. The framework was assessed on raw and segmented field‐captured images using three different SSL methods, Nearest Neighbor Contrastive Learning of Visual Representations (NNCLR), Bootstrap Your Own Latent, and Barlow Twins. SSL pre‐training was done on ResNet‐18 and ResNet‐50 models using all three SSL methods on the original RGB images and foreground segmented images. The performance of SSL pre‐training methods was evaluated using linear probing of SSL representations and end‐to‐end fine‐tuning approaches. The SSL‐pre‐trained convolutional neural network models were able to perform annotation‐efficient classification. NNCLR was the best performing SSL method for both linear and full model fine‐tuning. With just 5% annotated images, transfer learning with ImageNet initialization obtained 74% accuracy, whereas NNCLR achieved an improved classification accuracy of 79% for end‐to‐end fine‐tuning. Models created using SSL pre‐training consistently performed better, especially under very low annotation, and were robust to object class imbalances. These approaches help overcome annotation bottlenecks and are resource efficient. Core Ideas Insect pests cause significant damage to food production. Early detection and mitigation of insect pests are crucial in managing economic threshold level. We developed a self‐supervised learning (SSL) model to identify insect pests with minimal annotations. SSL models greatly improve the identification and classification tasks. Entropy‐masking‐based segmentation aids SSL effectiveness.
Journal Article
Automated trichome counting in soybean using advanced image‐processing techniques
2020
Premise Trichomes are hair‐like appendages extending from the plant epidermis. They serve many important biotic roles, including interference with herbivore movement. Characterizing the number, density, and distribution of trichomes can provide valuable insights on plant response to insect infestation and define the extent of plant defense capability. Automated trichome counting would speed up this research but poses several challenges, primarily because of the variability in coloration and the high occlusion of the trichomes. Methods and Results We developed a simplified method for image processing for automated and semi‐automated trichome counting. We illustrate this process using 30 leaves from 10 genotypes of soybean (Glycine max) differing in trichome abundance. We explored various heuristic image‐processing methods including thresholding and graph‐based algorithms to facilitate trichome counting. Of the two automated and two semi‐automated methods for trichome counting tested and with the help of regression analysis, the semi‐automated manually annotated trichome intersection curve method performed best, with an accuracy of close to 90% compared with the manually counted data. Conclusions We address trichome counting challenges including occlusion by combining image processing with human intervention to propose a semi‐automated method for trichome quantification. This provides new opportunities for the rapid and automated identification and quantification of trichomes, which has applications in a wide variety of disciplines.
Journal Article
How useful is active learning for image‐based plant phenotyping?
by
Fotouhi Ardakani, Fateme
,
Ganapathysubramanian, Baskar
,
Singh, Arti
in
Accuracy
,
Active learning
,
Bayesian analysis
2021
Deep learning models have been successfully deployed for a diverse array of image‐based plant phenotyping applications including disease detection and classification. However, successful deployment of supervised deep learning models requires large amount of labeled data, which is a significant challenge in plant sciences (and most biological) domain due to the inherent complexities. Specifically, data annotation is costly, laborious, time consuming and needs domain expertise for phenotyping tasks, especially for diseases. To overcome this challenge, active learning algorithms have been proposed to reduce the amount of labeling needed by deep learning models to achieve good predictive performance. Active learning methods work by adaptively suggesting samples to annotate using an acquisition function to achieve maximum (classification) performance under a fixed labeling budget. We report the performance of four different active learning methods, (1) Deep Bayesian Active Learning (DBAL), (2) Entropy, (3) Least Confidence, and (4) core‐set, with conventional random sampling‐based annotation for two vastly different image‐based classification datasets. The first image dataset consists of soybean [Glycine max L. (Merr.)] leaves belonging to eight different soybean stresses and a healthy class, and the second consists of nine different weed species from the field. For a fixed labeling budget, we observed that the classification performance of deep learning models using active learning based acquisition strategies is better than random sampling‐based acquisition for both datasets. The integration of active learning strategies for data annotation can help mitigate labelling challenges in the plant sciences applications particularly where resources dedicated to annotations are limited. Core Ideas Active learning methods reduce the amount of expert annotation needed in challenging image‐based plant classification tasks. Most acquisition functions built on uncertainty‐based sampling perform better than simple random sampling. However, random sampling is a good baseline for easy (for example, images under constant illumination conditions, i.e., less noisy data) classification tasks.
Journal Article
Plant Phenotyping with Limited Annotation: Doing More with Less
2022
Deep learning (DL) methods have transformed the way we extract plant traits – both under laboratory as well as field conditions. Extraction of plant phenotypes with DL can be a game changer in understanding and adapting crops for climate change. These extracted phenotypes can be linked with genotype, environment, management and their interactions to help farmers and breeders. Recent advancements in computation and sensor technology have enabled the cheap collection of high-resolution phenotype data across a large geographical area with high temporal resolution. Continuous increase in the amount of data collected and annotated has made it possible to apply deep learning algorithms successfully in a wide variety of challenging plant phenotyping tasks. Training a DL model typically requires the availability of copious amounts of annotated data; however, creating large-scale annotated dataset requires non-trivial efforts, time, and resources. This limitation has become a major bottleneck in deploying DL tools in practice. In this thesis, we will present a variety of ways in which one can circumvent the need for large annotated datasets for plant phenotyping. Using plant phenotyping project examples, we will illustrate transfer learning, active learning, self-supervised learning and domain adaption approaches. Utilizing these approaches can lower the barrier to deploying ML tools for plant science applications.
Dissertation
Self-Supervised Maize Kernel Classification and Segmentation for Embryo Identification
by
Baskar Ganapathysubramanian
,
Jubery, Talukder Z
,
Lübberstedt, Thomas
in
Classification
,
Corn
,
Deep learning
2022
Computer vision and deep learning (DL) techniques have succeeded in a wide range of diverse fields. Recently, these techniques have been successfully deployed in plant science applications to address food security, productivity, and environmental sustainability problems for a growing global population. However, training these DL models often necessitates the large-scale manual annotation of data which frequently becomes a tedious and time-and-resource-intensive process. Recent advances in self-supervised learning (SSL) methods have proven instrumental in overcoming these obstacles, using purely unlabeled datasets to pre-train DL models. Here, we implement the popular self-supervised contrastive learning methods of NNCLR (Nearest neighbor Contrastive Learning of visual Representations) and SimCLR (Simple framework for Contrastive Learning of visual Representations) for the classification of spatial orientation and segmentation of embryos of maize kernels. Maize kernels are imaged using a commercial high-throughput imaging system. This image data is often used in multiple downstream applications across both production and breeding applications, for instance, sorting for oil content based on segmenting and quantifying the scutellum's size and for classifying haploid and diploid kernels. We show that in both classification and segmentation problems, SSL techniques outperform their purely supervised transfer learning-based counterparts and are significantly more annotation efficient. Additionally, we show that a single SSL pre-trained model can be efficiently finetuned for both classification and segmentation, indicating good transferability across multiple downstream applications. Segmentation models with SSL-pretrained backbones produce DICE similarity coefficients of 0.81, higher than the 0.78 and 0.73 of those with ImageNet-pretrained and randomly initialized backbones, respectively. We observe that finetuning classification and segmentation models on as little as 1% annotation produces competitive results. These results show SSL provides a meaningful step forward in data efficiency with agricultural deep learning and computer vision.Competing Interest StatementThe authors have declared no competing interest.
Usefulness of interpretability methods to explain deep learning based plant stress phenotyping
by
Baskar Ganapathysubramanian
,
Singh, Arti
,
Nagasubramanian, Koushik
in
Classification
,
Color imagery
,
Decisions
2020
Deep learning techniques have been successfully deployed for automating plant stress identification and quantification. In recent years, there is a growing push towards training models that are interpretable -i.e. that justify their classification decisions by visually highlighting image features that were crucial for classification decisions. The expectation is that trained network models utilize image features that mimic visual cues used by plant pathologists. In this work, we compare some of the most popular interpretability methods: Saliency Maps, SmoothGrad, Guided Backpropogation, Deep Taylor Decomposition, Integrated Gradients, Layer-wise Relevance Propagation and Gradient times Input, for interpreting the deep learning model. We train a DenseNet-121 network for the classification of eight different soybean stresses (biotic and abiotic). Using a dataset consisting of 16,573 RGB images of healthy and stressed soybean leaflets captured under controlled conditions, we obtained an overall classification accuracy of 95.05 \\%. For a diverse subset of the test data, we compared the important features with those identified by a human expert. We observed that most interpretability methods identify the infected regions of the leaf as important features for some -- but not all -- of the correctly classified images. For some images, the output of the interpretability methods indicated that spurious feature correlations may have been used to correctly classify them. Although the output explanation maps of these interpretability methods may be different from each other for a given image, we advocate the use of these interpretability methods as `hypothesis generation' mechanisms that can drive scientific insight.