Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Reading LevelReading Level
-
Content TypeContent Type
-
YearFrom:-To:
-
More FiltersMore FiltersItem TypeIs Full-Text AvailableSubjectPublisherSourceDonorLanguagePlace of PublicationContributorsLocation
Done
Filters
Reset
111,233
result(s) for
"Biomedical data"
Sort by:
Automatic detection of lung cancer from biomedical data set using discrete AdaBoost optimized ensemble learning generalized neural networks
by
Al-Makhadmeh, Zafer
,
Mohamed, Shakeel P
,
Tolba Amr
in
Biomedical data
,
Datasets
,
Ensemble learning
2020
Today, most of the people are affected by lung cancer, mainly because of the genetic changes of the tissues in the lungs. Other factors such as smoking, alcohol, and exposure to dangerous gases can also be considered the contributory causes of lung cancer. Due to the serious consequences of lung cancer, the medical associations have been striving to diagnose cancer in its early stage of growth by applying the computer-aided diagnosis process. Although the CAD system at healthcare centers is able to diagnose lung cancer during its early stage of growth, the accuracy of cancer detection is difficult to achieve, mainly because of the overfitting of lung cancer features and the dimensionality of the feature set. Thus, this paper introduces the effective and optimized neural computing and soft computing techniques to minimize the difficulties and issues in the feature set. Initially, lung biomedical data were collected from the ELVIRA Biomedical Data Set Repository. The noise present in the data was eliminated by applying the bin smoothing normalization process. The minimum repetition and Wolf heuristic features were subsequently selected to minimize the dimensionality and complexity of the features. The selected lung features were analyzed using discrete AdaBoost optimized ensemble learning generalized neural networks, which successfully analyzed the biomedical lung data and classified the normal and abnormal features with great effectiveness. The efficiency of the system was then evaluated using MATLAB experimental setup in terms of error rate, precision, recall, G-mean, F-measure, and prediction rate.
Journal Article
High-risk multimorbidity patterns on the road to cardiovascular mortality
by
Thurner, Stefan
,
Gyimesi, Michael
,
Haug, Nina
in
Aged
,
Beyond Big Data to new Biomedical and Health Data Science: moving to next century precision health
,
Biomedical data
2020
Background
Multimorbidity, the co-occurrence of two or more diseases in one patient, is a frequent phenomenon. Understanding how different diseases condition each other over the lifetime of a patient could significantly contribute to personalised prevention efforts. However, most of our current knowledge on the long-term development of the health of patients (their disease trajectories) is either confined to narrow time spans or specific (sets of) diseases. Here, we aim to identify decisive events that potentially determine the future disease progression of patients.
Methods
Health states of patients are described by algorithmically identified multimorbidity patterns (groups of included or excluded diseases) in a population-wide analysis of 9,000,000 patient histories of hospital diagnoses observed over 17 years. Over time, patients might acquire new diagnoses that change their health state; they describe a disease trajectory. We measure the age- and sex-specific risks for patients that they will acquire certain sets of diseases in the future depending on their current health state.
Results
In the present analysis, the population is described by a set of 132 different multimorbidity patterns. For elderly patients, we find 3 groups of multimorbidity patterns associated with low (yearly in-hospital mortality of 0.2–0.3%), medium (0.3–1%) and high in-hospital mortality (2–11%). We identify combinations of diseases that significantly increase the risk to reach the high-mortality health states in later life. For instance, in men (women) aged 50–59 diagnosed with diabetes and hypertension, the risk for moving into the high-mortality region within 1 year is increased by the factor of 1.96 ± 0.11 (2.60 ± 0.18) compared with all patients of the same age and sex, respectively, and by the factor of 2.09 ± 0.12 (3.04 ± 0.18) if additionally diagnosed with metabolic disorders.
Conclusions
Our approach can be used both to forecast future disease burdens, as well as to identify the critical events in the careers of patients which strongly determine their disease progression, therefore constituting targets for efficient prevention measures. We show that the risk for cardiovascular diseases increases significantly more in females than in males when diagnosed with diabetes, hypertension and metabolic disorders.
Journal Article
Soft computing techniques for biomedical data analysis: open issues and challenges
2023
In recent years, medical data analysis has become paramount in delivering accurate diagnoses for various diseases. The plethora of medical data sources, encompassing disease types, disease-related proteins, ligands for proteins, and molecular drug components, necessitates adopting effective disease analysis and diagnosis methods. Soft computing techniques, including swarm algorithms and machine learning (ML) methods, have emerged as superior approaches. While ML techniques such as classification and clustering have gained prominence, feature selection methods are crucial in extracting optimal features and reducing data dimensions. This review paper presents a comprehensive overview of soft computing techniques for tackling medical data problems through classifying and analyzing medical data. The focus lies mainly on the classification of medical data resources. A detailed examination of various techniques developed for classifying numerous diseases is provided. The review encompasses an in-depth exploration of multiple ML methods designed explicitly for disease detection and classification. Additionally, the review paper offers insights into the underlying biological disease mechanisms and highlights several medical and chemical databases that facilitate research in this field. Furthermore, the review paper outlines emerging trends and identifies the key challenges in biomedical data analysis. It sheds light on this research domain’s exciting possibilities and future directions. The enhanced understanding of soft computing techniques and their practical applications and limitations will contribute to advancing biomedical data analysis and support healthcare professionals in making accurate diagnoses.
Journal Article
A review of automatic selection methods for machine learning algorithms and hyper-parameter values
Machine learning studies automatic algorithms that improve themselves through experience. It is widely used for analyzing and extracting value from large biomedical data sets, or “big biomedical data,” advancing biomedical research, and improving healthcare. Before a machine learning model is trained, the user of a machine learning software tool typically must manually select a machine learning algorithm and set one or more model parameters termed hyper-parameters. The algorithm and hyper-parameter values used can greatly impact the resulting model’s performance, but their selection requires special expertise as well as many labor-intensive manual iterations. To make machine learning accessible to layman users with limited computing expertise, computer science researchers have proposed various automatic selection methods for algorithms and/or hyper-parameter values for a given supervised machine learning problem. This paper reviews these methods, identifies several of their limitations in the big biomedical data environment, and provides preliminary thoughts on how to address these limitations. These findings establish a foundation for future research on automatically selecting algorithms and hyper-parameter values for analyzing big biomedical data.
Journal Article
Biomedical Big Data Technologies, Applications, and Challenges for Precision Medicine: A Review
2024
The explosive growth of biomedical Big Data presents both significant opportunities and challenges in the realm of knowledge discovery and translational applications within precision medicine. Efficient management, analysis, and interpretation of big data can pave the way for groundbreaking advancements in precision medicine. However, the unprecedented strides in the automated collection of large‐scale molecular and clinical data have also introduced formidable challenges in terms of data analysis and interpretation, necessitating the development of novel computational approaches. Some potential challenges include the curse of dimensionality, data heterogeneity, missing data, class imbalance, and scalability issues. This overview article focuses on the recent progress and breakthroughs in the application of big data within precision medicine. Key aspects are summarized, including content, data sources, technologies, tools, challenges, and existing gaps. Nine fields—Datawarehouse and data management, electronic medical record, biomedical imaging informatics, Artificial intelligence‐aided surgical design and surgery optimization, omics data, health monitoring data, knowledge graph, public health informatics, and security and privacy—are discussed. The explosive growth of biomedical Big Data presents both significant opportunities and challenges in the realm of knowledge discovery and translational applications within precision medicine. Efficient management, analysis, and interpretation of big data can pave the way for groundbreaking advancements in precision medicine. The review focuses on the recent progress and breakthroughs in the application of big data within precision medicine.
Journal Article
‘Fit-for-purpose?’ – challenges and opportunities for applications of blockchain technology in the future of healthcare
by
Clauson, Kevin A.
,
Kuo, Tsung-Ting
,
Church, George
in
Beyond Big Data to new Biomedical and Health Data Science moving to next century precision health
,
Biomedical Technology - methods
,
Biomedical Technology - organization & administration
2019
Blockchain is a shared distributed digital ledger technology that can better facilitate data management, provenance and security, and has the potential to transform healthcare. Importantly, blockchain represents a data architecture, whose application goes far beyond Bitcoin – the cryptocurrency that relies on blockchain and has popularized the technology. In the health sector, blockchain is being aggressively explored by various stakeholders to optimize business processes, lower costs, improve patient outcomes, enhance compliance, and enable better use of healthcare-related data. However, critical in assessing whether blockchain can fulfill the hype of a technology characterized as ‘revolutionary’ and ‘disruptive’, is the need to ensure that blockchain design elements consider actual healthcare needs from the diverse perspectives of consumers, patients, providers, and regulators. In addition, answering the real needs of healthcare stakeholders, blockchain approaches must also be responsive to the unique challenges faced in healthcare compared to other sectors of the economy. In this sense, ensuring that a health blockchain is ‘fit-for-purpose’ is pivotal. This concept forms the basis for this article, where we share views from a multidisciplinary group of practitioners at the forefront of blockchain conceptualization, development, and deployment.
Journal Article
PWSC: a novel clustering method based on polynomial weight-adjusted sparse clustering for sparse biomedical data and its application in cancer subtyping
2023
Background
Clustering analysis is widely used to interpret biomedical data and uncover new knowledge and patterns. However, conventional clustering methods are not effective when dealing with sparse biomedical data. To overcome this limitation, we propose a hierarchical clustering method called polynomial weight-adjusted sparse clustering (PWSC).
Results
The PWSC algorithm adjusts feature weights using a polynomial function, redefines the distances between samples, and performs hierarchical clustering analysis based on these adjusted distances. Additionally, we incorporate a consensus clustering approach to determine the optimal number of classifications. This consensus approach utilizes relative change in the cumulative distribution function to identify the best number of clusters, resulting in more stable clustering results. Leveraging the PWSC algorithm, we successfully classified a cohort of gastric cancer patients, enabling categorization of patients carrying different types of altered genes. Further evaluation using Entropy showed a significant improvement (
p
= 2.905e−05), while using the Calinski–Harabasz index demonstrates a remarkable 100% improvement in the quality of the best classification compared to conventional algorithms. Similarly, significantly increased entropy (
p
= 0.0336) and comparable CHI, were observed when classifying another colorectal cancer cohort with microbial abundance. The above attempts in cancer subtyping demonstrate that PWSC is highly applicable to different types of biomedical data. To facilitate its application, we have developed a user-friendly tool that implements the PWSC algorithm, which canbe accessed at
http://pwsc.aiyimed.com/
.
Conclusions
PWSC addresses the limitations of conventional approaches when clustering sparse biomedical data. By adjusting feature weights and employing consensus clustering, we achieve improved clustering results compared to conventional methods. The PWSC algorithm provides a valuable tool for researchers in the field, enabling more accurate and stable clustering analysis. Its application can enhance our understanding of complex biological systems and contribute to advancements in various biomedical disciplines.
Journal Article
Data–Driven Multimodal Sleep Apnea Events Detection
A novel multimodal and bio–inspired approach to biomedical signal processing and classification is presented in the paper. This approach allows for an automatic semantic labeling (interpretation) of sleep apnea events based the proposed data–driven biomedical signal processing and classification. The presented signal processing and classification methods have been already successfully applied to real–time unimodal brainwaves (EEG only) decoding in brain–computer interfaces developed by the author. In the current project the very encouraging results are obtained using multimodal biomedical (brainwaves and peripheral physiological) signals in a unified processing approach allowing for the automatic semantic data description. The results thus support a hypothesis of the data–driven and bio–inspired signal processing approach validity for medical data semantic interpretation based on the sleep apnea events machine–learning–related classification.
Journal Article
The Venus score for the assessment of the quality and trustworthiness of biomedical datasets
by
Fabris, Alessandro
,
Chicco, Davide
,
Jurman, Giuseppe
in
Algorithms
,
Archives & records
,
Artificial intelligence
2025
Biomedical datasets are the mainstays of computational biology and health informatics projects, and can be found on multiple data platforms online or obtained from wet-lab biologists and physicians. The quality and the trustworthiness of these datasets, however, can sometimes be poor, producing bad results in turn, which can harm patients and data subjects. To address this problem, policy-makers, researchers, and consortia have proposed diverse regulations, guidelines, and scores to assess the quality and increase the reliability of datasets. Although generally useful, however, they are often incomplete and impractical. The guidelines of
Datasheets for Datasets
, in particular, are too numerous; the requirements of the
Kaggle Dataset Usability Score
focus on non-scientific requisites (for example, including a cover image); and the
European Union Artificial Intelligence Act
(EU AI Act) sets forth sparse and general data governance requirements, which we tailored to datasets for biomedical AI. Against this backdrop, we introduce our new Venus score to assess the data quality and trustworthiness of biomedical datasets. Our score ranges from 0 to 10 and consists of ten questions that anyone developing a bioinformatics, medical informatics, or cheminformatics dataset should answer before the release. In this study, we first describe the
EU AI Act
,
Datasheets for Datasets
, and the
Kaggle Dataset Usability Score
, presenting their requirements and their drawbacks. To do so, we reverse-engineer the weights of the influential Kaggle Score for the first time and report them in this study. We distill the most important data governance requirements into ten questions tailored to the biomedical domain, comprising the Venus score. We apply the Venus score to twelve datasets from multiple subdomains, including electronic health records, medical imaging, microarray and bulk RNA-seq gene expression, cheminformatics, physiologic electrogram signals, and medical text. Analyzing the results, we surface fine-grained strengths and weaknesses of popular datasets, as well as aggregate trends. Most notably, we find a widespread tendency to gloss over sources of data inaccuracy and noise, which may hinder the reliable exploitation of data and, consequently, research results. Overall, our results confirm the applicability and utility of the Venus score to assess the trustworthiness of biomedical data.
Journal Article