ejeai Open Access Journal

European Journal of Emerging Artificial Intelligence

eISSN: Applied
Publication Frequency : 2 Issues per year.

  • Peer Reviewed & International Journal
Table of Content
Issues (Year-wise)
Loading…

Open Access iconOpen Access

ARTICLE

BRIDGING THE GENERALIZATION GAP IN VISUAL REINFORCEMENT LEARNING: A THEORETICAL AND EMPIRICAL STUDY

1 Department of Computer Science, Universidad de Buenos Aires, Argentina
2 Department of Computer Engineering, King Saud University, Saudi Arabia

Citations: Loading…
ABSTRACT VIEWS: 37   |   FILE VIEWS: 28   |   PDF: 28   HTML: 0   OTHER: 0   |   TOTAL: 65
Views + Downloads (Last 90 days)
Cumulative % included

Abstract

Visual Reinforcement Learning (VRL) agents frequently suffer from a significant "generalization gap," exhibiting degraded performance when deployed in environments that subtly differ from their training conditions. This paper provides a comprehensive analysis of the factors contributing to this discrepancy, integrating theoretical insights with empirical evidence. We categorize and discuss various strategies employed to bridge this gap, including the pivotal roles of data augmentation, advanced representation learning techniques (such as self-supervised and invariant learning), regularization methods, domain randomization for sim-to-real transfer, and the integration of auxiliary tasks and structured policy approaches. Our findings underscore the importance of learning robust, invariant visual representations and the efficacy of exposing agents to diverse, augmented experiences. We highlight the ongoing challenges, particularly in quantifying and optimizing for true environmental invariance, and propose future research directions aimed at developing more adaptable and generalizable VRL systems capable of thriving in varied real-world scenarios.


Keywords

Reinforcement Learning, Visual Reinforcement Learning, Generalization, Data Augmentation

References

1. Agarwal, A., Hsu, D. J., Kale, S., Langford, J., Li, L., & Schapire, R. E. (2014). Taming the Monster: A Fast and Simple Algorithm for Contextual Bandits. In International Conference on Machine Learning.

2. Agarwal, R., Machado, M. C., Castro, P. S., & Bellemare, M. G. (2021). Contrastive Behavioral Similarity Embeddings for Generalization in Reinforcement Learning. In International Conference on Learning Representations.

3. Agrawal, S., & Goyal, N. (2012). Thompson Sampling for Contextual Bandits with Linear Payoffs. In International Conference on Machine Learning.

4. Bahl, S., Mukadam, M., Gupta, A. K., & Pathak, D. (2020). Neural Dynamic Policies for End-to-End Sensorimotor Learning. In Neural Information Processing Systems.

5. Bauer, M., & Mnih, A. (2021). Generalized Doubly Reparameterized Gradient Estimators. In International Conference on Machine Learning.


How to Cite

BRIDGING THE GENERALIZATION GAP IN VISUAL REINFORCEMENT LEARNING: A THEORETICAL AND EMPIRICAL STUDY. (2024). European Journal of Emerging Artificial Intelligence, 1(01), 17-36. https://parthenonfrontiers.com/index.php/ejeai/article/view/46

Related articles

Share Link