Asset Details
MbrlCatalogueTitleDetail
Do you wish to reserve the book?
In-Context Decision Transformer: Reinforcement Learning via Hierarchical Chain-of-Thought
by
Chen, Hechang
, Huang, Sili
, Yang, Bo
, Hu, Jifeng
, Sun, Lichao
in
Benchmarks
/ Context
/ Decision making
/ Transformers
2024
Hey, we have placed the reservation for you!
By the way, why not check out events that you can attend while you pick your title.
You are currently in the queue to collect this book. You will be notified once it is your turn to collect the book.
Oops! Something went wrong.
Looks like we were not able to place the reservation. Kindly try again later.
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
Do you wish to request the book?
In-Context Decision Transformer: Reinforcement Learning via Hierarchical Chain-of-Thought
by
Chen, Hechang
, Huang, Sili
, Yang, Bo
, Hu, Jifeng
, Sun, Lichao
in
Benchmarks
/ Context
/ Decision making
/ Transformers
2024
Please be aware that the book you have requested cannot be checked out. If you would like to checkout this book, you can reserve another copy
We have requested the book for you!
Your request is successful and it will be processed during the Library working hours. Please check the status of your request in My Requests.
Oops! Something went wrong.
Looks like we were not able to place your request. Kindly try again later.
In-Context Decision Transformer: Reinforcement Learning via Hierarchical Chain-of-Thought
Paper
In-Context Decision Transformer: Reinforcement Learning via Hierarchical Chain-of-Thought
2024
Request Book From Autostore
and Choose the Collection Method
Overview
In-context learning is a promising approach for offline reinforcement learning (RL) to handle online tasks, which can be achieved by providing task prompts. Recent works demonstrated that in-context RL could emerge with self-improvement in a trial-and-error manner when treating RL tasks as an across-episodic sequential prediction problem. Despite the self-improvement not requiring gradient updates, current works still suffer from high computational costs when the across-episodic sequence increases with task horizons. To this end, we propose an In-context Decision Transformer (IDT) to achieve self-improvement in a high-level trial-and-error manner. Specifically, IDT is inspired by the efficient hierarchical structure of human decision-making and thus reconstructs the sequence to consist of high-level decisions instead of low-level actions that interact with environments. As one high-level decision can guide multi-step low-level actions, IDT naturally avoids excessively long sequences and solves online tasks more efficiently. Experimental results show that IDT achieves state-of-the-art in long-horizon tasks over current in-context RL methods. In particular, the online evaluation time of our IDT is \\textbf{36\\(\\times\\)} times faster than baselines in the D4RL benchmark and \\textbf{27\\(\\times\\)} times faster in the Grid World benchmark.
Publisher
Cornell University Library, arXiv.org
Subject
This website uses cookies to ensure you get the best experience on our website.