Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
LLMs are Also Effective Embedding Models: An In-depth Overview
by
Shen, Tao
, Li, Zhen
, Hua, Kai
, Chongyang Tao
, Zhang, Junshuo
, Ma, Shuai
, Gao, Shen
, Hu, Wenpeng
, Tao, Zhengwei
in
Effectiveness
/ Efficiency
/ Embedding
/ Large language models
/ Modal data
/ Natural language processing
/ Production methods
/ Scaling laws
/ Tuning
2025
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
LLMs are Also Effective Embedding Models: An In-depth Overview
by
Shen, Tao
, Li, Zhen
, Hua, Kai
, Chongyang Tao
, Zhang, Junshuo
, Ma, Shuai
, Gao, Shen
, Hu, Wenpeng
, Tao, Zhengwei
in
Effectiveness
/ Efficiency
/ Embedding
/ Large language models
/ Modal data
/ Natural language processing
/ Production methods
/ Scaling laws
/ Tuning
2025
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
LLMs are Also Effective Embedding Models: An In-depth Overview
by
Shen, Tao
, Li, Zhen
, Hua, Kai
, Chongyang Tao
, Zhang, Junshuo
, Ma, Shuai
, Gao, Shen
, Hu, Wenpeng
, Tao, Zhengwei
in
Effectiveness
/ Efficiency
/ Embedding
/ Large language models
/ Modal data
/ Natural language processing
/ Production methods
/ Scaling laws
/ Tuning
2025
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
LLMs are Also Effective Embedding Models: An In-depth Overview
Paper
LLMs are Also Effective Embedding Models: An In-depth Overview
2025
Request Book From Autostore
and Choose the Collection Method
Overview
Large language models (LLMs) have revolutionized natural language processing by achieving state-of-the-art performance across various tasks. Recently, their effectiveness as embedding models has gained attention, marking a paradigm shift from traditional encoder-only models like ELMo and BERT to decoder-only, large-scale LLMs such as GPT, LLaMA, and Mistral. This survey provides an in-depth overview of this transition, beginning with foundational techniques before the LLM era, followed by LLM-based embedding models through two main strategies to derive embeddings from LLMs. 1) Direct prompting: We mainly discuss the prompt designs and the underlying rationale for deriving competitive embeddings. 2) Data-centric tuning: We cover extensive aspects that affect tuning an embedding model, including model architecture, training objectives, data constructions, etc. Upon the above, we also cover advanced methods for producing embeddings from longer texts, multilingual, code, cross-modal data, as well as reasoning-aware and other domain-specific scenarios. Furthermore, we discuss factors affecting choices of embedding models, such as performance/efficiency comparisons, dense vs sparse embeddings, pooling strategies, and scaling law. Lastly, the survey highlights the limitations and challenges in adapting LLMs for embeddings, including cross-task embedding quality, trade-offs between efficiency and accuracy, low-resource, long-context, data bias, robustness, etc. This survey serves as a valuable resource for researchers and practitioners by synthesizing current advancements, highlighting key challenges, and offering a comprehensive framework for future work aimed at enhancing the effectiveness and efficiency of LLMs as embedding models.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.