Catalogue Search | MBRL
Search Results Heading
Explore the vast range of titles available.
MBRLSearchResults
-
DisciplineDiscipline
-
Is Peer ReviewedIs Peer Reviewed
-
Series TitleSeries Title
-
Reading LevelReading Level
-
YearFrom:-To:
-
More FiltersMore FiltersContent TypeItem TypeIs Full-Text AvailableSubjectCountry Of PublicationPublisherSourceDonorLanguagePlace of PublicationContributorsLocation
Done
Filters
Reset
9,722
result(s) for
"autonomous learning"
Sort by:
On the Construction of Autonomous Learning Environment in Colleges and Universities under the Condition of Network Information Technology
In recent years, colleges and universities pay more and more attention to the autonomous learning of college students. Autonomous learning is not only conducive to the improvement of students' academic performance, but also the premise of individual lifelong learning and lifelong development. Information technology has made great changes in people's life and learning style. Autonomous learning based on information technology has become an important learning style. Information technology can provide a strong support environment for autonomous learning, but it will not naturally become a good support for autonomous learning. We must design it carefully to make it play the maximum effect. From the perspective of subject teachers, the author will study the construction of autonomous learning environment based on information technology from two aspects of theory and practice, and provide good support for students' autonomous learning to improve students' learning performance and cultivate students' information literacy.
Journal Article
The developmental organization of robot behavior
by
Grupen, Roderic A., author
in
Autonomous robots.
,
Cooperating objects (Computer systems)
,
Machine learning.
2023
\"This book explores the question of how robots might \"learn\" how to behave in novel environments, without having each possible action pre-programmed\"-- Provided by publisher.
Vehicle maintenance management based on machine learning in agricultural tractor engines
by
Benavides-Cevallos, Ignacio Bayardo
,
Hernández-Rueda, Erik Paul
,
Mafla-Yépez, Carlos Nolasco
in
Agricultural equipment
,
Agricultural vehicles
,
Algorithms
2023
The objective of this work is to use the autonomous learning methodology as a tool in vehicle maintenance management. In obtaining data, faults in the fuel supply system have been simulated, causing anomalies in the combustion process that are easily detectable by vibrations obtained from a sensor in the engine of an agricultural tractor. To train the classification algorithm, 4 engine states were used: BE (optimal state), MEF1, MEF2, MEF3 (simulated failures). The applied autonomous learning is of the supervised type, where the samples were initially characterized and labeled to create a database for the execution of the training. The results show that the training carried out within the classification algorithm has an efficiency greater than 90%, which indicates that the method used is applicable in the management of vehicle maintenance to predict failures in engine operation.
Journal Article
Explainable reinforcement learning for broad-XAI: a conceptual framework and survey
by
Vamplew, Peter
,
Cruz, Francisco
,
Dazeley, Richard
in
Algorithms
,
Artificial Intelligence
,
Communication
2023
Broad-XAI
moves away from interpreting individual decisions based on a single datum and aims to provide integrated explanations from multiple machine learning algorithms into a coherent explanation of an agent’s behaviour that is aligned to the communication needs of the explainee. Reinforcement Learning (RL) methods, we propose, provide a potential backbone for the cognitive model required for the development of Broad-XAI. RL represents a suite of approaches that have had increasing success in solving a range of sequential decision-making problems. However, these algorithms operate as black-box problem solvers, where they obfuscate their decision-making policy through a complex array of values and functions. EXplainable RL (XRL) aims to develop techniques to extract concepts from the agent’s: perception of the environment; intrinsic/extrinsic motivations/beliefs; Q-values, goals and objectives. This paper aims to introduce the Causal XRL Framework (CXF), that unifies the current XRL research and uses RL as a backbone to the development of Broad-XAI. CXF is designed to incorporate many standard RL extensions and integrated with external ontologies and communication facilities so that the agent can answer questions that explain outcomes its decisions. This paper aims to: establish XRL as a distinct branch of XAI; introduce a conceptual framework for XRL; review existing approaches explaining agent behaviour; and identify opportunities for future research. Finally, this paper discusses how additional information can be extracted and ultimately integrated into models of communication, facilitating the development of Broad-XAI.
Journal Article
Knowledge- and ambiguity-aware robot learning from corrective and evaluative feedback
by
Kober, Jens
,
Celemin, Carlos
in
Ambiguity
,
Artificial Intelligence
,
Computational Biology/Bioinformatics
2023
In order to deploy robots that could be adapted by non-expert users, interactive imitation learning (IIL) methods must be flexible regarding the interaction preferences of the teacher and avoid assumptions of perfect teachers (oracles), while considering they make mistakes influenced by diverse human factors. In this work, we propose an IIL method that improves the human–robot interaction for non-expert and imperfect teachers in two directions. First, uncertainty estimation is included to endow the agents with a lack of knowledge awareness (epistemic uncertainty) and demonstration ambiguity awareness (aleatoric uncertainty), such that the robot can request human input when it is deemed more necessary. Second, the proposed method enables the teachers to train with the flexibility of using corrective demonstrations, evaluative reinforcements, and implicit positive feedback. The experimental results show an improvement in learning convergence with respect to other learning methods when the agent learns from highly ambiguous teachers. Additionally, in a user study, it was found that the components of the proposed method improve the teaching experience and the data efficiency of the learning process.
Journal Article
AI apology: interactive multi-objective reinforcement learning for human-aligned AI
by
Cruz, Francisco
,
Dazeley, Richard
,
Vamplew, Peter
in
Alignment
,
Artificial Intelligence
,
Behavior
2023
For an Artificially Intelligent (AI) system to maintain alignment between human desires and its behaviour, it is important that the AI account for human preferences. This paper proposes and empirically evaluates the first approach to aligning agent behaviour to human preference via an
apologetic
framework. In practice, an apology may consist of an acknowledgement, an explanation and an intention for the improvement of future behaviour. We propose that such an apology, provided in response to recognition of undesirable behaviour, is one way in which an AI agent may both be transparent and trustworthy to a human user. Furthermore, that behavioural adaptation as part of apology is a viable approach to correct against undesirable behaviours. The Act-Assess-Apologise framework potentially could address both the practical and social needs of a human user, to recognise and make reparations against prior undesirable behaviour and adjust for the future. Applied to a dual-auxiliary impact minimisation problem, the apologetic agent had a near perfect determination and apology provision accuracy in several non-trivial configurations. The agent subsequently demonstrated behaviour alignment with success that included up to complete avoidance of the impacts described by these objectives in some scenarios.
Journal Article
Hierarchical goals contextualize local reward decomposition explanations
by
Heintz, Fredrik
,
Rietz, Finn
,
Stork, Johannes A.
in
Artificial Intelligence
,
Computational Biology/Bioinformatics
,
Computational Science and Engineering
2023
One-step reinforcement learning explanation methods account for individual actions but fail to consider the agent’s future behavior, which can make their interpretation ambiguous. We propose to address this limitation by providing hierarchical goals as context for one-step explanations. By considering the current hierarchical goal as a context, one-step explanations can be interpreted with higher certainty, as the agent’s future behavior is more predictable. We combine reward decomposition with hierarchical reinforcement learning into a novel explainable reinforcement learning framework, which yields more interpretable, goal-contextualized one-step explanations. With a qualitative analysis of one-step reward decomposition explanations, we first show that their interpretability is indeed limited in scenarios with multiple, different optimal policies—a characteristic shared by other one-step explanation methods. Then, we show that our framework retains high interpretability in such cases, as the hierarchical goal can be considered as context for the explanation. To the best of our knowledge, our work is the first to investigate hierarchical goals not as an explanation directly but as additional context for one-step reinforcement learning explanations.
Journal Article
Proxemic behavior in navigation tasks using reinforcement learning
by
Fernandes, Bruno
,
Millán-Arias, Cristian
,
Cruz, Francisco
in
Agents (artificial intelligence)
,
Animal cognition
,
Artificial Intelligence
2023
Human interaction starts with a person approaching another one, respecting their personal space to prevent uncomfortable feelings. Spatial behavior, called proxemics, allows defining an acceptable distance so that the interaction process begins appropriately. In recent decades, human-agent interaction has been an area of interest for researchers, where it is proposed that artificial agents naturally interact with people. Thus, new alternatives are needed to allow optimal communication, avoiding humans feeling uncomfortable. Several works consider proxemic behavior with cognitive agents, where human-robot interaction techniques and machine learning are implemented. However, it is assumed that the personal space is fixed and known in advance, and the agent is only expected to make an optimal trajectory toward the person. In this work, we focus on studying the behavior of a reinforcement learning agent in a proxemic-based environment. Experiments were carried out implementing a grid-world problem and a continuous simulated robotic approaching environment. These environments assume that there is an issuer agent that provides non-conformity information. Our results suggest that the agent can identify regions where the issuer feels uncomfortable and find the best path to approach the issuer. The results obtained highlight the usefulness of reinforcement learning in order to identify proxemic regions.
Journal Article