Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
14 result(s) for "Gurushanth, Keerthi"
Sort by:
Quantitative Autofluorescence Imaging of Oral Mucosa and Lesions: A Proof-of-Concept Study
Background/Objectives: This study aimed to quantitatively assess site-specific mean autofluorescence intensity across normal oral mucosal subsites and to evaluate the effectiveness of Autofluorescence Imaging (AFI) as an adjunct tool for distinguishing benign lesions, OPMDs, and oral cancers by comparing lesion intensity with anatomically matched healthy subsites. Methods: This observational study employed dual-mode imaging, comprising paired White Light Imaging (WLI) and AFI, captured from different oral cavity subsites using a smartphone-based point-of-care device. The Region of Interest (ROI) was annotated on WLI and automatically mapped to the corresponding AFI for both normal mucosa and lesions. WLI and AFI images were separated into their constituent red, green, and blue (RGB) channels, and AFI intensity was quantified via ImageJ. Results: A total of 1380 dual-mode images were acquired from 86 healthy participants. AFI intensities were comparable across most oral subsites, except for the lateral and ventral tongue. The lateral border showed the lowest fluorescence (Green channel-GC: 68.12 ± 28.27; Blue channel-BC: 25.29 ± 7.93), whereas the ventral tongue showed the highest (GC: 98.89 ± 42.22; BC: 37.08 ± 11.04; both p < 0.001). Among 611 lesions, predominantly from the buccal mucosa, AFI intensity declined progressively with increasing disease severity. Homogeneous leukoplakia (n = 149; GC: 38.62 ± 25.05; BC: 21.60 ± 9.50), non-homogeneous leukoplakia (n = 25; GC: 30.42 ± 18.66; BC: 18.25 ± 7.17) and oral cancer (n = 21; GC: 23.39 ± 15.53; BC: 15.82 ± 7.15; all p < 0.001) showed markedly reduced fluorescence, while benign lesions (n: 44; GC: 66.99 ± 30.88; BC: 32.01 ± 13.62) exhibited intermediate intensities, supporting AFI’s discriminative potential. Conclusions: This phase-1, proof-of-concept study highlights subsite-specific variations in autofluorescence intensity within healthy oral mucosa, providing an essential baseline for objective interpretation of lesion-associated fluorescence changes. AFI has the potential to be used as a non-invasive adjunct for monitoring OPMDs. Further validation in larger and more diverse cohorts is required before clinical implementation.
Comparison of gray values of cone-beam computed tomography with hounsfield units of multislice computed tomography: An in vitro study
Purpose: Hounsfield unit (HU) provides a quantitative evaluation of bone density. The assessment of bone density is essential for successful treatment plan. Although, multislice computed tomography (MSCT) is considered as gold standard in evaluating bone density, cone-beam computed tomography (CBCT) is frequently used in dentomaxillofacial imaging due to lower radiation dose, less complex device, and images with satisfactory resolution. Aims and Objectives: The aim of this study is to determine and compare the gray value and HU value of hypodense and hyperdense structures on CBCT and MSCT, respectively. The study also evaluated and compared the gray values in different field of views within CBCT. Materials and Methods: A total of 20 dry human mandibles were obtained. The gray values and HU values of hypodense structures (extraction socket, inferior alveolar canal, and mental foramen) and hyperdense structures (enamel, cancellous, and cortical bone) were evaluated and compared between CBCT and MSCT images, respectively. The obtained data were statistically analyzed. Statistical Analysis: One-way analyses of variance, ANOVA F-test. Results: The gray value for hypodense structures in large volume CBCT scans resembled the HU value. The study showed statistically significant difference (P < 0.001) in gray values for all the hyperdense structures in CBCT when compared to HU values of MSCT scans. Conclusion: The gray value for hypodense structures in large volume CBCT scan was more reliable and analogous to HU value in MSCT. The determination of grey values in CBCT may not be as accurate as HU value in CT for hyperdense structures.
Determination of Proximity of Mandibular Third Molar to Mandibular Canal Using Panoramic Radiography and Cone-beam Computed Tomography
Objectives: Position of inferior alveolar canal with respect to an impacted third molar reveals certain radiographic signs, but three-dimensional relationship to the canal can be provided with cone-beam computed tomography (CBCT). The purpose of this study was to determine which radiographic signs on panoramic radiography indicate a true relationship on CBCT. Materials and Methods: Forty samples with signs or symptoms of impacted mandibular third molar and panoramic radiograph showing signs of a close relationship with the mandibular canal as described by Félez-Gutiérrez et al. were included in the study and subjected to CBCT. Radiographic signs on panoramic radiography were compared with the relationship on CBCT. Statistical analysis was done using Chi-square test. Results: Twenty-one samples (52.5%) showed darkening of the apex, which was the most frequent type of radiographic sign of a close relationship on panoramic radiography. Twenty-three samples (57.5%) revealed a true relationship on CBCT. Darkening of the apex and narrowing of the canal were the signs most frequently associated with a true relationship. On CBCT, coronal and axial sections better predicted a true relationship. Conclusion: This study showed that the presence of any of the radiographic signs cannot definitely predict a true relationship; however, the presence of a close sign on panoramic radiography is often associated with a true relationship to the canal.
An asymptomatic radiolucent lesion in posterior mandible: A case report
Radiolucent mandibular lesions are commonly evident on head and neck imaging and present a diagnostic dilemma for the radiologist. These may represent a broad spectrum of lesions arising from both odontogenic and nonodontogenic structures. Furthermore, few radiolucent lesions are often identified as incidental lesions by the radiologist on imaging performed for different reasons. Location of the lesion, borders, internal structure, and its effect on surrounding structures are the key points to narrow the differential diagnosis. Imaging is essential not only for the diagnosis of lesions, but also to guide therapy and monitor the treatment response. Here is a case report on traumatic bone cyst that presented as an asymptomatic radiolucent lesion in right posterior mandible and was discovered incidentally on routine radiograph. The lesion was diagnosed based on patient′s anamnesis and radiographic examination. This paper aims at discussing the differential diagnosis, various radiological characteristics, and their prediction in prognosis of the lesion.
Interpretable deep learning approach for oral cancer classification using guided attention inference network
Significance: Convolutional neural networks (CNNs) show the potential for automated classification of different cancer lesions. However, their lack of interpretability and explainability makes CNNs less than understandable. Furthermore, CNNs may incorrectly concentrate on other areas surrounding the salient object, rather than the network’s attention focusing directly on the object to be recognized, as the network has no incentive to focus solely on the correct subjects to be detected. This inhibits the reliability of CNNs, especially for biomedical applications. Aim: Develop a deep learning training approach that could provide understandability to its predictions and directly guide the network to concentrate its attention and accurately delineate cancerous regions of the image. Approach: We utilized Selvaraju et al.’s gradient-weighted class activation mapping to inject interpretability and explainability into CNNs. We adopted a two-stage training process with data augmentation techniques and Li et al.’s guided attention inference network (GAIN) to train images captured using our customized mobile oral screening devices. The GAIN architecture consists of three streams of network training: classification stream, attention mining stream, and bounding box stream. By adopting the GAIN training architecture, we jointly optimized the classification and segmentation accuracy of our CNN by treating these attention maps as reliable priors to develop attention maps with more complete and accurate segmentation. Results: The network’s attention map will help us to actively understand what the network is focusing on and looking at during its decision-making process. The results also show that the proposed method could guide the trained neural network to highlight and focus its attention on the correct lesion areas in the images when making a decision, rather than focusing its attention on relevant yet incorrect regions. Conclusions: We demonstrate the effectiveness of our approach for more interpretable and reliable oral potentially malignant lesion and malignant lesion classification.
Mobile-based oral cancer classification for point-of-care screening
Significance: Oral cancer is among the most common cancers globally, especially in low- and middle-income countries. Early detection is the most effective way to reduce the mortality rate. Deep learning-based cancer image classification models usually need to be hosted on a computing server. However, internet connection is unreliable for screening in low-resource settings. Aim: To develop a mobile-based dual-mode image classification method and customized Android application for point-of-care oral cancer detection. Approach: The dataset used in our study was captured among 5025 patients with our customized dual-modality mobile oral screening devices. We trained an efficient network MobileNet with focal loss and converted the model into TensorFlow Lite format. The finalized lite format model is ∼16.3  MB and ideal for smartphone platform operation. We have developed an Android smartphone application in an easy-to-use format that implements the mobile-based dual-modality image classification approach to distinguish oral potentially malignant and malignant images from normal/benign images. Results: We investigated the accuracy and running speed on a cost-effective smartphone computing platform. It takes ∼300  ms to process one image pair with the Moto G5 Android smartphone. We tested the proposed method on a standalone dataset and achieved 81% accuracy for distinguishing normal/benign lesions from clinically suspicious lesions, using a gold standard of clinical impression based on the review of images by oral specialists. Conclusions: Our study demonstrates the effectiveness of a mobile-based approach for oral cancer screening in low-resource settings.
Classification of imbalanced oral cancer image data from high-risk population
Significance: Early detection of oral cancer is vital for high-risk patients, and machine learning-based automatic classification is ideal for disease screening. However, current datasets collected from high-risk populations are unbalanced and often have detrimental effects on the performance of classification. Aim: To reduce the class bias caused by data imbalance. Approach: We collected 3851 polarized white light cheek mucosa images using our customized oral cancer screening device. We use weight balancing, data augmentation, undersampling, focal loss, and ensemble methods to improve the neural network performance of oral cancer image classification with the imbalanced multi-class datasets captured from high-risk populations during oral cancer screening in low-resource settings. Results: By applying both data-level and algorithm-level approaches to the deep learning training process, the performance of the minority classes, which were difficult to distinguish at the beginning, has been improved. The accuracy of “premalignancy” class is also increased, which is ideal for screening applications. Conclusions: Experimental results show that the class bias induced by imbalanced oral cancer image datasets could be reduced using both data- and algorithm-level methods. Our study may provide an important basis for helping understand the influence of unbalanced datasets on oral cancer deep learning classifiers and how to mitigate.
Small form factor, flexible, dual-modality handheld probe for smartphone-based, point-of-care oral and oropharyngeal cancer screening
Oral cancer is a growing health issue in low- and middle-income countries due to betel quid, tobacco, and alcohol use and in younger populations of middle- and high-income communities due to the prevalence of human papillomavirus. The described point-of-care, smartphone-based intraoral probe enables autofluorescence imaging and polarized white light imaging in a compact geometry through the use of a USB-connected camera module. The small size and flexible imaging head improves on previous intraoral probe designs and allows imaging the cheek pockets, tonsils, and base of tongue, the areas of greatest risk for both causes of oral cancer. Cloud-based remote specialist and convolutional neural network clinical diagnosis allow for both remote community and home use. The device is characterized and preliminary field-testing data are shared.
Exploring uncertainty measures in convolutional neural network for semantic segmentation of oral cancer images
Oral cancer is one of the most prevalent cancers, especially in middle- and low-income countries such as India. Automatic segmentation of oral cancer images can improve the diagnostic workflow, which is a significant task in oral cancer image analysis. Despite the remarkable success of deep-learning networks in medical segmentation, they rarely provide uncertainty quantification for their output. We aim to estimate uncertainty in a deep-learning approach to semantic segmentation of oral cancer images and to improve the accuracy and reliability of predictions. This work introduced a UNet-based Bayesian deep-learning (BDL) model to segment potentially malignant and malignant lesion areas in the oral cavity. The model can quantify uncertainty in predictions. We also developed an efficient model that increased the inference speed, which is almost six times smaller and two times faster (inference speed) than the original UNet. The dataset in this study was collected using our customized screening platform and was annotated by oral oncology specialists. The proposed approach achieved good segmentation performance as well as good uncertainty estimation performance. In the experiments, we observed an improvement in pixel accuracy and mean intersection over union by removing uncertain pixels. This result reflects that the model provided less accurate predictions in uncertain areas that may need more attention and further inspection. The experiments also showed that with some performance compromises, the efficient model reduced computation time and model size, which expands the potential for implementation on portable devices used in resource-limited settings. Our study demonstrates the UNet-based BDL model not only can perform potentially malignant and malignant oral lesion segmentation, but also can provide informative pixel-level uncertainty estimation. With this extra uncertainty information, the accuracy and reliability of the model’s prediction can be improved.
Inter-observer agreement among specialists in the diagnosis of oral potentially malignant disorders and oral cancer using store-and-forward technology
ObjectivesOral cancer is a leading cause of morbidity and mortality. Screening and mobile Health (mHealth)-based approach facilitates early detection remotely in a resource-limited settings. Recent advances in eHealth technology have enabled remote monitoring and triage to detect oral cancer in its early stages. Although studies have been conducted to evaluate the diagnostic efficacy of remote specialists, to our knowledge, no studies have been conducted to evaluate the consistency of remote specialists. The aim of this study was to evaluate interobserver agreement between specialists through telemedicine systems in real-world settings using store-and-forward technology.Materials and methodsThe two remote specialists independently diagnosed clinical images (n=822) from image archives. The onsite specialist diagnosed the same participants using conventional visual examination, which was tabulated. The diagnostic accuracy of two remote specialists was compared with that of the onsite specialist. Images that were confirmed histopathologically were compared with the onsite diagnoses and the two remote specialists.ResultsThere was moderate agreement (k= 0.682) between two remote specialists and (k= 0.629) between the onsite specialist and two remote specialists in the diagnosis of oral lesions. The sensitivity and specificity of remote specialist 1 were 92.7% and 83.3%, respectively, and those of remote specialist 2 were 95.8% and 60%, respectively, each compared with histopathology.ConclusionThe diagnostic accuracy of the two remote specialists was optimal, suggesting that “store and forward” technology and telehealth can be an effective tool for triage and monitoring of patients.Clinical relevanceTelemedicine is a good tool for triage and enables faster patient care in real-world settings.