Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning Dataset
by
Moshkov, Ivan
, Toshniwal, Shubham
, Gitman, Daria
, Jia, Fei
, Gitman, Igor
, Narenthiran, Sean
in
Datasets
/ Large language models
/ Mathematics education
/ Skills
/ Tuning
2024
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning Dataset
by
Moshkov, Ivan
, Toshniwal, Shubham
, Gitman, Daria
, Jia, Fei
, Gitman, Igor
, Narenthiran, Sean
in
Datasets
/ Large language models
/ Mathematics education
/ Skills
/ Tuning
2024
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning Dataset
Paper
OpenMathInstruct-1: A 1.8 Million Math Instruction Tuning Dataset
2024
Request Book From Autostore
and Choose the Collection Method
Overview
Recent work has shown the immense potential of synthetically generated datasets for training large language models (LLMs), especially for acquiring targeted skills. Current large-scale math instruction tuning datasets such as MetaMathQA (Yu et al., 2024) and MAmmoTH (Yue et al., 2024) are constructed using outputs from closed-source LLMs with commercially restrictive licenses. A key reason limiting the use of open-source LLMs in these data generation pipelines has been the wide gap between the mathematical skills of the best closed-source LLMs, such as GPT-4, and the best open-source LLMs. Building on the recent progress in open-source LLMs, our proposed prompting novelty, and some brute-force scaling, we construct OpenMathInstruct-1, a math instruction tuning dataset with 1.8M problem-solution pairs. The dataset is constructed by synthesizing code-interpreter solutions for GSM8K and MATH, two popular math reasoning benchmarks, using the recently released and permissively licensed Mixtral model. Our best model, OpenMath-CodeLlama-70B, trained on a subset of OpenMathInstruct-1, achieves a score of 84.6% on GSM8K and 50.7% on MATH, which is competitive with the best gpt-distilled models. We release our code, models, and the OpenMathInstruct-1 dataset under a commercially permissive license.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.