Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Language
      Language
      Clear All
      Language
  • Subject
      Subject
      Clear All
      Subject
  • Item Type
      Item Type
      Clear All
      Item Type
  • Discipline
      Discipline
      Clear All
      Discipline
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
22 result(s) for "Rallo, Robert"
Sort by:
Molecular structural dataset of lignin macromolecule elucidating experimental structural compositions
Lignin is one of the most abundant biopolymers in nature and has great potential to be transformed into high-value chemicals. However, the limited availability of molecular structure data hinders its potential industrial applications. Herein, we present the Lignin Structural (LGS) Dataset that includes the molecular structure of milled wood lignin focusing on two major monomeric units (coniferyl and syringyl), and the six most common interunit linkages (phenylpropane β-aryl ether, resinol, phenylcoumaran, biphenyl, dibenzodioxocin, and diaryl ether). The dataset constitutes a unique resource that covers a part of lignin’s chemical space characterized by polymer chains with lengths in the range of 3 to 25 monomer units. Structural data were generated using a sequence-controlled polymer generation approach that was calibrated to match experimental lignin properties. The LGS dataset includes 60 K newly generated lignin structures that match with high accuracy (~90%) the experimentally determined structural compositions available in the literature. The LGS dataset is a valuable resource to advance lignin chemistry research, including computational simulation approaches and predictive modelling.Measurement(s)molecular structureTechnology Type(s)Computer ModelingFactor Type(s)monomer ratio • bond frequency • degree of polymerizationSample Characteristic - Organismconiferous (softwood) • deciduous (hardwood)
Preparing for the next pandemic via transfer learning from existing diseases with hierarchical multi-modal BERT: a study on COVID-19 outcome prediction
Developing prediction models for emerging infectious diseases from relatively small numbers of cases is a critical need for improving pandemic preparedness. Using COVID-19 as an exemplar, we propose a transfer learning methodology for developing predictive models from multi-modal electronic healthcare records by leveraging information from more prevalent diseases with shared clinical characteristics. Our novel hierarchical, multi-modal model ( T R A N S MED ) integrates baseline risk factors from the natural language processing of clinical notes at admission, time-series measurements of biomarkers obtained from laboratory tests, and discrete diagnostic, procedure and drug codes. We demonstrate the alignment of T R A N S MED ’s predictions with well-established clinical knowledge about COVID-19 through univariate and multivariate risk factor driven sub-cohort analysis. T R A N S MED ’s superior performance over state-of-the-art methods shows that leveraging patient data across modalities and transferring prior knowledge from similar disorders is critical for accurate prediction of patient outcomes, and this approach may serve as an important tool in the early response to future pandemics.
Automated Phenotype Recognition for Zebrafish Embryo Based In Vivo High Throughput Toxicity Screening of Engineered Nano-Materials
A phenotype recognition model was developed for high throughput screening (HTS) of engineered Nano-Materials (eNMs) toxicity using zebrafish embryo developmental response classified, from automatically captured images and without manual manipulation of zebrafish positioning, by three basic phenotypes (i.e., hatched, unhatched, and dead). The recognition model was built with a set of vectorial descriptors providing image color and texture information. The best performing model was attained with three image descriptors (color histogram, representative color, and color layout) identified as most suitable from an initial pool of six descriptors. This model had an average recognition accuracy of 97.40±0.95% in a 10-fold cross-validation and 93.75% in a stress test of low quality zebrafish images. The present work has shown that a phenotyping model can be developed with accurate recognition ability suitable for zebrafish-based HTS assays. Although the present methodology was successfully demonstrated for only three basic zebrafish embryonic phenotypes, it can be readily adapted to incorporate more subtle phenotypes.
An ISA-TAB-Nano based data collection framework to support data-driven modelling of nanotoxicology
Analysis of trends in nanotoxicology data and the development of data driven models for nanotoxicity is facilitated by the reporting of data using a standardised electronic format. ISA-TAB-Nano has been proposed as such a format. However, in order to build useful datasets according to this format, a variety of issues has to be addressed. These issues include questions regarding exactly which (meta)data to report and how to report them. The current article discusses some of the challenges associated with the use of ISA-TAB-Nano and presents a set of resources designed to facilitate the manual creation of ISA-TAB-Nano datasets from the nanotoxicology literature. These resources were developed within the context of the NanoPUZZLES EU project and include data collection templates, corresponding business rules that extend the generic ISA-TAB-Nano specification as well as Python code to facilitate parsing and integration of these datasets within other nanoinformatics resources. The use of these resources is illustrated by a “Toy Dataset” presented in the Supporting Information. The strengths and weaknesses of the resources are discussed along with possible future developments.
CATMoS: Collaborative Acute Toxicity Modeling Suite
Humans are exposed to tens of thousands of chemical substances that need to be assessed for their potential toxicity. Acute systemic toxicity testing serves as the basis for regulatory hazard classification, labeling, and risk management. However, it is cost- and time-prohibitive to evaluate all new and existing chemicals using traditional rodent acute toxicity tests. models built using existing data facilitate rapid acute toxicity predictions without using animals. The U.S. Interagency Coordinating Committee on the Validation of Alternative Methods (ICCVAM) Acute Toxicity Workgroup organized an international collaboration to develop models for predicting acute oral toxicity based on five different end points: Lethal Dose 50 ( value, U.S. Environmental Protection Agency hazard (four) categories, Globally Harmonized System for Classification and Labeling hazard (five) categories, very toxic chemicals [ ( )], and nontoxic chemicals ( ). An acute oral toxicity data inventory for 11,992 chemicals was compiled, split into training and evaluation sets, and made available to 35 participating international research groups that submitted a total of 139 predictive models. Predictions that fell within the applicability domains of the submitted models were evaluated using external validation sets. These were then combined into consensus models to leverage strengths of individual approaches. The resulting consensus predictions, which leverage the collective strengths of each individual model, form the Collaborative Acute Toxicity Modeling Suite (CATMoS). CATMoS demonstrated high performance in terms of accuracy and robustness when compared with results. CATMoS is being evaluated by regulatory agencies for its utility and applicability as a potential replacement for rat acute oral toxicity studies. CATMoS predictions for more than 800,000 chemicals have been made available via the National Toxicology Program's Integrated Chemical Environment tools and data sets (ice.ntp.niehs.nih.gov). The models are also implemented in a free, standalone, open-source tool, OPERA, which allows predictions of new and untested chemicals to be made. https://doi.org/10.1289/EHP8495.
A US perspective on closing the carbon cycle to defossilize difficult-to-electrify segments of our economy
Electrification to reduce or eliminate greenhouse gas emissions is essential to mitigate climate change. However, a substantial portion of our manufacturing and transportation infrastructure will be difficult to electrify and/or will continue to use carbon as a key component, including areas in aviation, heavy-duty and marine transportation, and the chemical industry. In this Roadmap, we explore how multidisciplinary approaches will enable us to close the carbon cycle and create a circular economy by defossilizing these difficult-to-electrify areas and those that will continue to need carbon. We discuss two approaches for this: developing carbon alternatives and improving our ability to reuse carbon, enabled by separations. Furthermore, we posit that co-design and use-driven fundamental science are essential to reach aggressive greenhouse gas reduction targets. To achieve net-zero carbon emissions, we must close the carbon cycle for industries that are difficult to electrify. Developing the needed science to provide carbon alternatives and non-fossil carbon will accelerate advances towards defossilization.
Fractal Dimension Calculation for Big Data Using Box Locality Index
The box - counting approach for fractal dimension calculation is scaled up for big data using a data structure named box locality index (BLI). The BLI is constructed as key-value pairs with the key indexing the location of a “box” (i.e., a grid cell on the multi-dimensional space) and the value counting the number of data points inside the box (i.e., “box occupancy”). Such a key-value pair structure of BLI significantly simplifies the traditionally used hierarchical structure and encodes only necessary information required by the box - counting approach for fractal dimension calculation. Moreover, as the box occupancies (i.e., the values) associated with the same index (i.e., the key) are aggregatable, the BLI grants the box - counting approach the needed scalability for fractal dimension calculation of big data using distributed computing techniques (e.g., MapReduce and Spark). Taking the advantage of the BLI, MapReduce and Spark methods for fractal dimension calculation of big data are developed, which conduct box - counting for each grid level as a cascade of MapReduce/Spark jobs in a bottom-up fashion. In an empirical validation, the MapReduce and Spark methods demonstrated good effectiveness and efficiency in fractal calculation of a big synthetic dataset. In summary, this work provides an efficient solution for estimating the intrinsic dimension of big data, which is essential for many machine learning methods and data analytics including feature selection and dimensionality reduction.
Prediction of the Q-e parameters from structures of transfer chain agents
The empirical parameters of copolymerization Q-e have been examined as an endpoint for establishing the quantitative structure – property relationships (QSPRs). The possibility to build up QSPR for these parameters is demonstrated for 22 transfer chain agents. Data for 20 taken in the literature and two were investigated in direct experiment. The statistical qualities of the models for parameter e together with the negative decimal logarithm of Q × 10 −4 (pQ) are quite good. The mechanistic interpretation for these models are suggested and discussed.