Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
101 result(s) for "Remondini, Daniel"
Sort by:
Comparison between 16S rRNA and shotgun sequencing data for the taxonomic characterization of the gut microbiota
In this paper we compared taxonomic results obtained by metataxonomics (16S rRNA gene sequencing) and metagenomics (whole shotgun metagenomic sequencing) to investigate their reliability for bacteria profiling, studying the chicken gut as a model system. The experimental conditions included two compartments of gastrointestinal tracts and two sampling times. We compared the relative abundance distributions obtained with the two sequencing strategies and then tested their capability to distinguish the experimental conditions. The results showed that 16S rRNA gene sequencing detects only part of the gut microbiota community revealed by shotgun sequencing. Specifically, when a sufficient number of reads is available, Shotgun sequencing has more power to identify less abundant taxa than 16S sequencing. Finally, we showed that the less abundant genera detected only by shotgun sequencing are biologically meaningful, being able to discriminate between the experimental conditions as much as the more abundant genera detected by both sequencing strategies.
Methods for the integration of multi-omics data: mathematical aspects
Background Methods for the integrative analysis of multi-omics data are required to draw a more complete and accurate picture of the dynamics of molecular systems. The complexity of biological systems, the technological limits, the large number of biological variables and the relatively low number of biological samples make the analysis of multi-omics datasets a non-trivial problem. Results and Conclusions We review the most advanced strategies for integrating multi-omics datasets, focusing on mathematical and methodological aspects.
AC amplification gain in organic electrochemical transistors for impedance-based single cell sensors
Research on electrolyte-gated and organic electrochemical transistor (OECT) architectures is motivated by the prospect of a highly biocompatible interface capable of amplifying bioelectronic signals at the site of detection. Despite many demonstrations in these directions, a quantitative model for OECTs as impedance biosensors is still lacking. We overcome this issue by introducing a model experiment where we simulate the detection of a single cell by the impedance sensing of a dielectric microparticle. The highly reproducible experiment allows us to study the impact of transistor geometry and operation conditions on device sensitivity. With the data we rationalize a mathematical model that provides clear guidelines for the optimization of OECTs as single cell sensors, and we verify the quantitative predictions in an in-vitro experiment. In the optimized geometry, the OECT-based impedance sensor allows to record single cell adhesion and detachment transients, showing a maximum gain of 20.2±0.9 dB with respect to a single electrode-based impedance sensor. The authors develop a quantitative description of alternating current amplification gain in organic electrochemical transistors. The findings are applied to achieve detection of single glioblastoma cell adhesion with 20 dB gain compared to microelectrodes.
Weighted Multiplex Networks
One of the most important challenges in network science is to quantify the information encoded in complex network structures. Disentangling randomness from organizational principles is even more demanding when networks have a multiplex nature. Multiplex networks are multilayer systems of [Formula: see text] nodes that can be linked in multiple interacting and co-evolving layers. In these networks, relevant information might not be captured if the single layers were analyzed separately. Here we demonstrate that such partial analysis of layers fails to capture significant correlations between weights and topology of complex multiplex networks. To this end, we study two weighted multiplex co-authorship and citation networks involving the authors included in the American Physical Society. We show that in these networks weights are strongly correlated with multiplex structure, and provide empirical evidence in favor of the advantage of studying weighted measures of multiplex networks, such as multistrength and the inverse multiparticipation ratio. Finally, we introduce a theoretical framework based on the entropy of multiplex ensembles to quantify the information stored in multiplex networks that would remain undetected if the single layers were analyzed in isolation.
Leveraging complex network features improves vaccine stance classification
The widespread use of social media allows unprecedented ways to monitor opinions and stances regarding critical public health issues globally. Advanced Natural Language processing algorithms are being used routinely to extract information and classify vaccination hesitancy or stance. However, communication on online social networks such as Twitter (now X) is carried by short messages, the meaning of which can be difficult to understand in the absence of context. Therefore, in this study we propose the use of complex-network features extracted from the social network to integrate and enhance text-based Deep Learning models. Leveraging a dataset of about 20 million Italian language posts (of which about 7000 were manually annotated), we showed how the integration of text and network features improves vaccine stance classification, especially for the most polarized classes. Additionally, network features overperformed text features in a dataset collected a year after model training, possibly indicating how the social network changes more slowly than the trending words or topics.
Clusters of science and health related Twitter users become more isolated during the COVID-19 pandemic
COVID-19 represents the most severe global crisis to date whose public conversation can be studied in real time. To do so, we use a data set of over 350 million tweets and retweets posted by over 26 million English speaking Twitter users from January 13 to June 7, 2020. We characterize the retweet network to identify spontaneous clustering of users and the evolution of their interaction over time in relation to the pandemic’s emergence. We identify several stable clusters (super-communities), and are able to link them to international groups mainly involved in science and health topics, national elites, and political actors. The science- and health-related super-community received disproportionate attention early on during the pandemic, and was leading the discussion at the time. However, as the pandemic unfolded, the attention shifted towards both national elites and political actors, paralleled by the introduction of country-specific containment measures and the growing politicization of the debate. Scientific super-community remained present in the discussion, but experienced less reach and became more isolated within the network. Overall, the emerging network communities are characterized by an increased self-amplification and polarization. This makes it generally harder for information from international health organizations or scientific authorities to directly reach a broad audience through Twitter for prolonged time. These results may have implications for information dissemination along the unfolding of long-term events like epidemic diseases on a world-wide scale.
Sediment core analysis using artificial intelligence
Subsurface stratigraphic modeling is crucial for a variety of environmental, societal, and economic challenges. However, the need for specific sedimentological skills in sediment core analysis may constitute a limitation. Methods based on Machine Learning and Deep Learning can play a central role in automatizing this time-consuming procedure. In this work, using a robust dataset of high-resolution digital images from continuous sediment cores of Holocene age that reflect a wide spectrum of continental to shallow-marine depositional environments, we outline a novel deep-learning-based approach to perform automatic semantic segmentation directly on core images, leveraging the power of convolutional neural networks. To optimize the interpretation process and maximize scientific value, we use six sedimentary facies associations as target classes in lieu of ineffective classification methods based uniquely on lithology. We propose an automated model that can rapidly characterize sediment cores, allowing immediate guidance for stratigraphic correlation and subsurface reconstructions.
A network approach for low dimensional signatures from high throughput data
One of the main objectives of high-throughput genomics studies is to obtain a low-dimensional set of observables—a signature—for sample classification purposes (diagnosis, prognosis, stratification). Biological data, such as gene or protein expression, are commonly characterized by an up/down regulation behavior, for which discriminant-based methods could perform with high accuracy and easy interpretability. To obtain the most out of these methods features selection is even more critical, but it is known to be a NP-hard problem, and thus most feature selection approaches focuses on one feature at the time (k-best, Sequential Feature Selection, recursive feature elimination). We propose DNetPRO, Discriminant Analysis with Network PROcessing , a supervised network-based signature identification method. This method implements a network-based heuristic to generate one or more signatures out of the best performing feature pairs. The algorithm is easily scalable, allowing efficient computing for high number of observables ( 10 3 – 10 5 ). We show applications on real high-throughput genomic datasets in which our method outperforms existing results, or is compatible with them but with a smaller number of selected features. Moreover, the geometrical simplicity of the resulting class-separation surfaces allows a clearer interpretation of the obtained signatures in comparison to nonlinear classification models.
Language models learn to represent antigenic properties of human influenza A(H3) virus
Given that influenza vaccine effectiveness depends on a good antigenic match between the vaccine and circulating viruses, it is important to assess the antigenic properties of newly emerging variants continuously. With the increasing application of real-time pathogen genomic surveillance, a key question is if antigenic properties can reliably be predicted from influenza virus genomic information. Based on validated linked datasets of influenza virus genomic and wet lab experimental results, in silico models may be of use to learn to predict immune escape of variants of interest starting from the protein sequence only. In this study, we compared several machine-learning methods to reconstruct antigenic map coordinates for HA1 protein sequences of influenza A(H3N2) virus, to rank substitutions responsible for major antigenic changes, and to recognize variants with novel antigenic properties that may warrant future vaccine updates. Methods based on deep learning language models (BiLSTM and ProtBERT) and more classical approaches based solely on genetic distances and physicochemical properties of amino acid sequences had comparable performances over the coarser features of the map, but the first two performed better over fine-grained features like single amino acid-driven antigenic change and in silico deep mutational scanning experiments to rank the substitutions with the largest impact on antigenic properties. Given that the best performing model that produces protein embeddings is agnostic to the specific pathogen, the presented approach may be applicable to other pathogens.
Optimized pipeline of MuTect and GATK tools to improve the detection of somatic single nucleotide polymorphisms in whole-exome sequencing data
Background Detecting somatic mutations in whole exome sequencing data of cancer samples has become a popular approach for profiling cancer development, progression and chemotherapy resistance. Several studies have proposed software packages, filters and parametrizations. However, many research groups reported low concordance among different methods. We aimed to develop a pipeline which detects a wide range of single nucleotide mutations with high validation rates. We combined two standard tools – Genome Analysis Toolkit (GATK) and MuTect – to create the GATK-LOD N method. As proof of principle, we applied our pipeline to exome sequencing data of hematological (Acute Myeloid and Acute Lymphoblastic Leukemias) and solid (Gastrointestinal Stromal Tumor and Lung Adenocarcinoma) tumors. We performed experiments on simulated data to test the sensitivity and specificity of our pipeline. Results The software MuTect presented the highest validation rate (90 %) for mutation detection, but limited number of somatic mutations detected. The GATK detected a high number of mutations but with low specificity. The GATK-LOD N increased the performance of the GATK variant detection (from 5 of 14 to 3 of 4 confirmed variants), while preserving mutations not detected by MuTect. However, GATK-LOD N filtered more variants in the hematological samples than in the solid tumors. Experiments in simulated data demonstrated that GATK-LOD N increased both specificity and sensitivity of GATK results. Conclusion We presented a pipeline that detects a wide range of somatic single nucleotide variants, with good validation rates, from exome sequencing data of cancer samples. We also showed the advantage of combining standard algorithms to create the GATK-LOD N method, that increased specificity and sensitivity of GATK results. This pipeline can be helpful in discovery studies aimed to profile the somatic mutational landscape of cancer genomes.