Search Results Heading

MBRLSearchResults

mbrl.module.common.modules.added.book.to.shelf
Title added to your shelf!
View what I already have on My Shelf.
Oops! Something went wrong.
Oops! Something went wrong.
While trying to add the title to your shelf something went wrong :( Kindly try again later!
Are you sure you want to remove the book from the shelf?
Oops! Something went wrong.
Oops! Something went wrong.
While trying to remove the title from your shelf something went wrong :( Kindly try again later!
    Done
    Filters
    Reset
  • Discipline
      Discipline
      Clear All
      Discipline
  • Is Peer Reviewed
      Is Peer Reviewed
      Clear All
      Is Peer Reviewed
  • Item Type
      Item Type
      Clear All
      Item Type
  • Subject
      Subject
      Clear All
      Subject
  • Year
      Year
      Clear All
      From:
      -
      To:
  • More Filters
2 result(s) for "multirobot cooperative planning"
Sort by:
Conclusions and Future Directions
This chapter concludes the book. The book identifies a few fundamental problems in multi‐robot coordination and proposes solutions to handle these problems by extending the traditional evolutionary algorithm (EA) and multi‐agent Q‐learning. It provides the preliminaries of Reinforcement Learning and EA in view of the multirobot coordination. The book proposes useful characteristic properties for exploration of the team‐goal and joint action selection in multi‐agent system. It also proposes a novel Consensus Q‐learning algorithm for multirobot cooperative planning. The book introduces a novel approach to correlated Q‐learning and subsequent multi‐robot planning. It also introduces a novel approach for efficiently employing both Imperialist Competitive Algorithm and Firefly Algorithm to develop a hybrid algorithm with an aim to utilize the composite benefits of the explorative and exploitative capabilities of both ancestor algorithms.
Consensus Q-Learning for Multi-agent Cooperative Planning
This chapter proposes consensus‐based multi‐agent Q‐learning (MAQL) to address the bottleneck of the optimal equilibrium selection among multiple types. It briefly introduces the adaption mechanism of single agent QL and the state‐of‐the‐art equilibrium‐based MAQL algorithms. Then the cooperative control problem employing PGs mainly focusing upon the consensus problem is briefly discussed. The chapter also proposes a novel consensus QL (CoQL). Subsequently, a consensus‐based multirobot cooperative planning algorithm is proposed. Then two experiments are presented. The first experiment is designed to study the relative performance of the CoQL over the reference algorithms. Another experiment is framed to study the relative performance of the consensus‐based planning algorithm over the reference algorithms, considering multi‐robot stick carrying problem as a benchmark in terms of state‐transitions required to complete the task.