Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
by
Zhu, Zhenkai
, Crous, Carl
, Tomala, Alex
, Akbulut, Canfer
, Steiner, David
, Krishna, Kalpesh
, Assael, Yannis
, Zheng, Ce
, Cobon-Kerr, James
, Jafari, Mohsen
, Cui, Albert
, Chi, Ed
, Abbas, Zaheer
, Dasgupta, Ishita
, Dotiwalla, Xerxes
, Altun, Yasemin
, Kassner, Nora
, Dukkipati, Nandita
, Kazawa, Hideto
, Slone, Ambrose
, Schrittwieser, Julian
, Borsos, Zalan
, Chen, Warren Weilun
, Sheng, XiangHai
, Matan Eyal
, Rajkumar, Samuel
, Goker Erdogan
, Xu, Jun
, Quinn, Michael
, Borgeaud, Sebastian
, Lazaridou, Angeliki
, Dragan, Anca
, Jazayeri, Sadegh
, Kouridi, Christina
, Liu, Rosanne
, Bapna, Ankur
, Trinh, Trieu
, Vashisht, Harsha
, Iqbal, Shariq
, Agrawal, Priyanka
, Popovici, Dan
, Andreev, Alek
, Raphael Lopez Kaufman
, Giang, Minh
, Paterson, Kim
, Carvajal, Gabriel
, Ding, Wen
, Cortes, Mario
, Orban, Andras
, Hauth, Anja
, Shabat, Nir
, Laskin, Michael
, Thacker, Phoebe
, Solomon, Kim
, Tucker, George
, Munkhdalai, Tsendsuren
, Chadwick, Martin
, Schucher, Nathan
, Liechty, Tyler
, Carey, C J
, Bahargam, Sanaz
, Senter, Evan
, Baddepudi, Anirudh
, Mao-Jones, Justin
, Zelle, Dustin
, Siddhartha Reddy Jonnalagadda
, Trdin, Nejc
, Chen, Wei
, Martens, James
, Nicosia, Massimo
, Vi
in
Context
/ Documents
/ Large language models
/ Retrieval
2024
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
by
Zhu, Zhenkai
, Crous, Carl
, Tomala, Alex
, Akbulut, Canfer
, Steiner, David
, Krishna, Kalpesh
, Assael, Yannis
, Zheng, Ce
, Cobon-Kerr, James
, Jafari, Mohsen
, Cui, Albert
, Chi, Ed
, Abbas, Zaheer
, Dasgupta, Ishita
, Dotiwalla, Xerxes
, Altun, Yasemin
, Kassner, Nora
, Dukkipati, Nandita
, Kazawa, Hideto
, Slone, Ambrose
, Schrittwieser, Julian
, Borsos, Zalan
, Chen, Warren Weilun
, Sheng, XiangHai
, Matan Eyal
, Rajkumar, Samuel
, Goker Erdogan
, Xu, Jun
, Quinn, Michael
, Borgeaud, Sebastian
, Lazaridou, Angeliki
, Dragan, Anca
, Jazayeri, Sadegh
, Kouridi, Christina
, Liu, Rosanne
, Bapna, Ankur
, Trinh, Trieu
, Vashisht, Harsha
, Iqbal, Shariq
, Agrawal, Priyanka
, Popovici, Dan
, Andreev, Alek
, Raphael Lopez Kaufman
, Giang, Minh
, Paterson, Kim
, Carvajal, Gabriel
, Ding, Wen
, Cortes, Mario
, Orban, Andras
, Hauth, Anja
, Shabat, Nir
, Laskin, Michael
, Thacker, Phoebe
, Solomon, Kim
, Tucker, George
, Munkhdalai, Tsendsuren
, Chadwick, Martin
, Schucher, Nathan
, Liechty, Tyler
, Carey, C J
, Bahargam, Sanaz
, Senter, Evan
, Baddepudi, Anirudh
, Mao-Jones, Justin
, Zelle, Dustin
, Siddhartha Reddy Jonnalagadda
, Trdin, Nejc
, Chen, Wei
, Martens, James
, Nicosia, Massimo
, Vi
in
Context
/ Documents
/ Large language models
/ Retrieval
2024
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
by
Zhu, Zhenkai
, Crous, Carl
, Tomala, Alex
, Akbulut, Canfer
, Steiner, David
, Krishna, Kalpesh
, Assael, Yannis
, Zheng, Ce
, Cobon-Kerr, James
, Jafari, Mohsen
, Cui, Albert
, Chi, Ed
, Abbas, Zaheer
, Dasgupta, Ishita
, Dotiwalla, Xerxes
, Altun, Yasemin
, Kassner, Nora
, Dukkipati, Nandita
, Kazawa, Hideto
, Slone, Ambrose
, Schrittwieser, Julian
, Borsos, Zalan
, Chen, Warren Weilun
, Sheng, XiangHai
, Matan Eyal
, Rajkumar, Samuel
, Goker Erdogan
, Xu, Jun
, Quinn, Michael
, Borgeaud, Sebastian
, Lazaridou, Angeliki
, Dragan, Anca
, Jazayeri, Sadegh
, Kouridi, Christina
, Liu, Rosanne
, Bapna, Ankur
, Trinh, Trieu
, Vashisht, Harsha
, Iqbal, Shariq
, Agrawal, Priyanka
, Popovici, Dan
, Andreev, Alek
, Raphael Lopez Kaufman
, Giang, Minh
, Paterson, Kim
, Carvajal, Gabriel
, Ding, Wen
, Cortes, Mario
, Orban, Andras
, Hauth, Anja
, Shabat, Nir
, Laskin, Michael
, Thacker, Phoebe
, Solomon, Kim
, Tucker, George
, Munkhdalai, Tsendsuren
, Chadwick, Martin
, Schucher, Nathan
, Liechty, Tyler
, Carey, C J
, Bahargam, Sanaz
, Senter, Evan
, Baddepudi, Anirudh
, Mao-Jones, Justin
, Zelle, Dustin
, Siddhartha Reddy Jonnalagadda
, Trdin, Nejc
, Chen, Wei
, Martens, James
, Nicosia, Massimo
, Vi
in
Context
/ Documents
/ Large language models
/ Retrieval
2024
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
Paper
Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context
2024
Request Book From Autostore
and Choose the Collection Method
Overview
In this report, we introduce the Gemini 1.5 family of models, representing the next generation of highly compute-efficient multimodal models capable of recalling and reasoning over fine-grained information from millions of tokens of context, including multiple long documents and hours of video and audio. The family includes two new models: (1) an updated Gemini 1.5 Pro, which exceeds the February version on the great majority of capabilities and benchmarks; (2) Gemini 1.5 Flash, a more lightweight variant designed for efficiency with minimal regression in quality. Gemini 1.5 models achieve near-perfect recall on long-context retrieval tasks across modalities, improve the state-of-the-art in long-document QA, long-video QA and long-context ASR, and match or surpass Gemini 1.0 Ultra's state-of-the-art performance across a broad set of benchmarks. Studying the limits of Gemini 1.5's long-context ability, we find continued improvement in next-token prediction and near-perfect retrieval (>99%) up to at least 10M tokens, a generational leap over existing models such as Claude 3.0 (200k) and GPT-4 Turbo (128k). Finally, we highlight real-world use cases, such as Gemini 1.5 collaborating with professionals on completing their tasks achieving 26 to 75% time savings across 10 different job categories, as well as surprising new capabilities of large language models at the frontier; when given a grammar manual for Kalamang, a language with fewer than 200 speakers worldwide, the model learns to translate English to Kalamang at a similar level to a person who learned from the same content.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.