Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Item TypeItem Type
-
SubjectSubject
-
YearFrom:-To:
-
More FiltersMore FiltersSourceLanguage
Done
Filters
Reset
23
result(s) for
"Toshniwal, Shubham"
Sort by:
A territorial scale analysis based on a map-oriented database for the out-of-plane seismic behaviour of masonry structures
2024
Since 1986, earthquakes have occurred in East Groningen, but most buildings made of unreinforced masonry have not been designed for earthquakes, which must now withstand magnitude 4 earthquakes. This has created an urgent need to assess large amounts of buildings in a fast but reliable manner. The out-of-plane behaviour is important for seismic assessments of unreinforced masonry buildings. Although the most accurate analysis method to determine the out-of-plane response of such walls is non-linear time-history analysis (NLTH), non-linear kinematic analysis (NLKA) provides a simple, fast but still reliable solution due to the computational difficulties of NLTH for structures constructed of unreinforced masonry. In this paper, the out-of-plane behaviours of masonry structures are up-scaled from a component scale to a provincial scale in a multi-scale manner. A map-oriented database is established to describe both local behaviours of walls and global behaviours of a province. The out-of-plane assessment by non-linear kinematic analysis (NLKA) is automated via the database without further calculations after the static analysis. The database provides a solid guidance to determine which detailed assessment methods will be adopted with limited data before a FEM model is built.
Journal Article
ASSESSMENT OF PULMONARY HYPERTENSION BY 2D ECHO AND DOPPLER IN PATIENT OF CHRONIC KIDNEY DISEASE
by
Pawar, Dinesh
,
Toshniwal, Shubham
,
Humaney, Nalini
in
Cardiovascular disease
,
Catheters
,
Fistula
2022
Chronic kidney disease (CKD) is major public health concern that leads to renal failure,cardiovascular disease(CVD),and early mortality.One significant cardiovascular consequence of CKD is pulmonary hypertension (PH). It is small arterial disease involving pulmonary vasculature that raises pulmonary vascular resistance (PVR) and in turn pulmonary arterial pressure (PAP).
Journal Article
ASSESSMENT OF PULMONARY HYPERTENSION BY 2D ECHO AND DOPPLER IN PATIENT OF CHRONIC KIDNEY DISEASE
by
Pawar, Dinesh
,
Toshniwal, Shubham
,
Humaney, Nalini
in
Cardiovascular disease
,
Catheters
,
Fistula
2022
Chronic kidney disease (CKD) is major public health concern that leads to renal failure,cardiovascular disease(CVD),and early mortality.One significant cardiovascular consequence of CKD is pulmonary hypertension (PH). It is small arterial disease involving pulmonary vasculature that raises pulmonary vascular resistance (PVR) and in turn pulmonary arterial pressure (PAP).
Journal Article
Efficient and Interpretable Neural Models for Entity Tracking
2022
What would it take for a natural language model to understand a novel, such as The Lord of the Rings? Among other things, such a model must be able to: (a) identify and record new characters (entities) and their attributes as they are introduced in the text, and (b) identify subsequent references to the characters previously introduced and update their attributes. This problem of entity tracking is essential for language understanding, and thus, useful for a wide array of downstream applications in NLP such as question-answering, summarization. In this thesis, we focus on two key problems in relation to facilitating the use of entity tracking models: (i) scaling entity tracking models to long documents, such as a novel, and (ii) integrating entity tracking into language models. Applying language technologies to long documents has garnered interest recently, but computational constraints are a significant bottleneck in scaling up current methods. In this thesis, we argue that computationally efficient entity tracking models can be developed by representing entities with rich, fixed-dimensional vector representations derived from pretrained language models, and by exploiting the ephemeral nature of entities. We also argue for the integration of entity tracking into language models as it will allow for: (i) wider application given the current ubiquitous use of pretrained language models in NLP applications, and (ii) easier adoption since it is much easier to swap in a new pretrained language model than to integrate a separate standalone entity tracking model.
Code Pretraining Improves Entity Tracking Abilities of Language Models
by
Toshniwal, Shubham
,
Schuster, Sebastian
,
Kim, Najoung
in
Alignment
,
Structured data
,
Tracking
2024
Recent work has provided indirect evidence that pretraining language models on code improves the ability of models to track state changes of discourse entities expressed in natural language. In this work, we systematically test this claim by comparing pairs of language models on their entity tracking performance. Critically, the pairs consist of base models and models trained on top of these base models with additional code data. We extend this analysis to additionally examine the effect of math training, another highly structured data type, and alignment tuning, an important step for enhancing the usability of models. We find clear evidence that models additionally trained on large amounts of code outperform the base models. On the other hand, we find no consistent benefit of additional math training or alignment tuning across various model families.
IdentifyMe: A Challenging Long-Context Mention Resolution Benchmark
by
Manikantan, Kawshik
,
Toshniwal, Shubham
,
Tapaswi, Makarand
in
Benchmarks
,
Multiple choice
,
Performance evaluation
2024
Recent evaluations of LLMs on coreference resolution have revealed that traditional output formats and evaluation metrics do not fully capture the models' referential understanding. To address this, we introduce IdentifyMe, a new benchmark for mention resolution presented in a multiple-choice question (MCQ) format, commonly used for evaluating LLMs. IdentifyMe features long narratives and employs heuristics to exclude easily identifiable mentions, creating a more challenging task. The benchmark also consists of a curated mixture of different mention types and corresponding entities, allowing for a fine-grained analysis of model performance. We evaluate both closed- and open source LLMs on IdentifyMe and observe a significant performance gap (20-30%) between the state-of-the-art sub-10B open models vs. closed ones. We observe that pronominal mentions, which have limited surface information, are typically much harder for models to resolve than nominal mentions. Additionally, we find that LLMs often confuse entities when their mentions overlap in nested structures. The highest-scoring model, GPT-4o, achieves 81.9% accuracy, highlighting the strong referential capabilities of state-of-the-art LLMs while also indicating room for further improvement.
Major Entity Identification: A Generalizable Alternative to Coreference Resolution
2024
The limited generalization of coreference resolution (CR) models has been a major bottleneck in the task's broad application. Prior work has identified annotation differences, especially for mention detection, as one of the main reasons for the generalization gap and proposed using additional annotated target domain data. Rather than relying on this additional annotation, we propose an alternative referential task, Major Entity Identification (MEI), where we: (a) assume the target entities to be specified in the input, and (b) limit the task to only the frequent entities. Through extensive experiments, we demonstrate that MEI models generalize well across domains on multiple datasets with supervised models and LLM-based few-shot prompting. Additionally, MEI fits the classification framework, which enables the use of robust and intuitive classification-based metrics. Finally, MEI is also of practical use as it allows a user to search for all mentions of a particular entity or a group of entities of interest.
Learning to Reason and Memorize with Self-Notes
2023
Large language models have been shown to struggle with multi-step reasoning, and do not retain previous reasoning steps for future use. We propose a simple method for solving both of these problems by allowing the model to take Self-Notes. Unlike recent chain-of-thought or scratchpad approaches, the model can deviate from the input context at any time to explicitly think and write down its thoughts. This allows the model to perform reasoning on the fly as it reads the context and even integrate previous reasoning steps, thus enhancing its memory with useful information and enabling multi-step reasoning. Experiments across a wide variety of tasks demonstrate that our method can outperform chain-of-thought and scratchpad methods by taking Self-Notes that interleave the input text.
OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning Dataset
by
Moshkov, Ivan
,
Toshniwal, Shubham
,
Gitman, Daria
in
Datasets
,
Large language models
,
Mathematics education
2024
Recent work has shown the immense potential of synthetically generated datasets for training large language models (LLMs), especially for acquiring targeted skills. Current large-scale math instruction tuning datasets such as MetaMathQA (Yu et al., 2024) and MAmmoTH (Yue et al., 2024) are constructed using outputs from closed-source LLMs with commercially restrictive licenses. A key reason limiting the use of open-source LLMs in these data generation pipelines has been the wide gap between the mathematical skills of the best closed-source LLMs, such as GPT-4, and the best open-source LLMs. Building on the recent progress in open-source LLMs, our proposed prompting novelty, and some brute-force scaling, we construct OpenMathInstruct-1, a math instruction tuning dataset with 1.8M problem-solution pairs. The dataset is constructed by synthesizing code-interpreter solutions for GSM8K and MATH, two popular math reasoning benchmarks, using the recently released and permissively licensed Mixtral model. Our best model, OpenMath-CodeLlama-70B, trained on a subset of OpenMathInstruct-1, achieves a score of 84.6% on GSM8K and 50.7% on MATH, which is competitive with the best gpt-distilled models. We release our code, models, and the OpenMathInstruct-1 dataset under a commercially permissive license.
OpenMathInstruct-2: Accelerating AI for Math with Massive Open-Source Instruction Data
2024
Mathematical reasoning continues to be a critical challenge in large language model (LLM) development with significant interest. However, most of the cutting-edge progress in mathematical reasoning with LLMs has become \\emph{closed-source} due to lack of access to training data. This lack of data access limits researchers from understanding the impact of different choices for synthesizing and utilizing the data. With the goal of creating a high-quality finetuning (SFT) dataset for math reasoning, we conduct careful ablation experiments on data synthesis using the recently released \\texttt{Llama3.1} family of models. Our experiments show that: (a) solution format matters, with excessively verbose solutions proving detrimental to SFT performance, (b) data generated by a strong teacher outperforms \\emph{on-policy} data generated by a weak student model, (c) SFT is robust to low-quality solutions, allowing for imprecise data filtering, and (d) question diversity is crucial for achieving data scaling gains. Based on these insights, we create the OpenMathInstruct-2 dataset, which consists of 14M question-solution pairs (\\(\\approx\\) 600K unique questions), making it nearly eight times larger than the previous largest open-source math reasoning dataset. Finetuning the \\texttt{Llama-3.1-8B-Base} using OpenMathInstruct-2 outperforms \\texttt{Llama3.1-8B-Instruct} on MATH by an absolute 15.9\\% (51.9\\% \\(\\rightarrow\\) 67.8\\%). Finally, to accelerate the open-source efforts, we release the code, the finetuned models, and the OpenMathInstruct-2 dataset under a commercially permissive license.