Deep Reinforcement Learning: Bridging Theory and Practice

ISAIM Special Session on Deep Reinforcement Learning


Submission deadline: Jan 1 2024
4-page paper


Bahia Mar Fort Lauderdale Beach, USA, Grand View and Harbor Lights Rooms

Date: Jan 8 and 9, 2024


<
Motivation

As a special session of the International Symposium on Artificial Intelligence and Mathematics (ISAIM 2024, Fort Lauderdale, FL, January 8–10, 2024), we are hosting the "Deep Reinforcement Learning: Bridging Theory and Practice" workshop. Deep Reinforcement Learning (DRL) is at the forefront of AI research, driving advancements in areas from robotics to finance. While there has been significant progress in reinforcement learning both from the perspective of theoretical advancements and from the perspective of empirical results, these communities have been largely disjoint. As in other fields, a bridging of both the theoretical and empirical viewpoint would be hugely beneficial to both communities and raise interesting and pertinent research questions! The goal of this workshop is to bridge this gap and foster an understanding and appreciation of the latest advancements in deep reinforcement learning (DRL) by emphasizing the connection between theoretical foundations and practical implementations, and to address the existing challenges facing the field. By bringing together researchers doing research on theoretical results, algorithm design and finally application to downstream realistic domains, we will build collaborations and insights that can be fruitful for both parties. While the potential of DRL is vast, there are inherent challenges that need to be addressed, both in theory and in practice. This workshop aims to provide participants with a comprehensive understanding of these challenges, backed by hands-on experiences and in-depth discussions.

Our workshop aims to bring together researchers studying the theory, and practice of deep RL, and RL algorithms more broadly to consider questions such as (but not limited to):

  • Sample Efficiency: DRL algorithms often require vast amounts of data to train. What strategies can make techniques more data-efficient, and practical for real world problems?
  • Exploration vs. Exploitation: Balancing the need to explore the environment and exploit known strategies remains a key challenge. How can we optimize this balance to yield more practical algorithms?
  • Transfer Learning: How can we leverage knowledge from one domain to enhance performance in another?
  • Stability and Convergence: With deep neural networks at their core, DRL algorithms can be prone to instability. What strategies can ensure stable learning and convergence?
  • Data and Scaling: Scaling has proven to be extremely effective in domains such as NLP and vision. How do these insights extend to deep RL algorithms?
  • Human-in-the-loop RL: RL algorithms will be deployed around people. How can these algorithms take insights from cognitive science/behavioral methods and apply them to develop provably safe/efficient DRL methods?
  • Real-world Application Barriers: Deploying DRL in real-world scenarios poses its own set of challenges, from safety concerns to integration issues. Our workshop will provide insights into navigating these challenges effectively.
More broadly, we encourage researchers from any field related to reinforcement learning whether it be theory or practice to contribute their work! We look forward to a broad-ranging and fun discussion. The workshop is open to all - researchers in academia or industry, AI practitioners, graduate students, and industry professionals interested in the cutting-edge developments of Deep Reinforcement Learning and its challenges.


Speakers

Bo Dai

Georgia Tech

Aviral Kumar

Carnegie Mellon University

Wen Sun

Cornell University

Sanjiban Choudhury

Cornell University

Animesh Garg

Georgia Institute of Technology

Zhuoran Yang

Yale University

Chi Jin

Princeton University

Reuth Mirsky

Bar-Ilan University


Schedule (Jan 8)

Time (GMT-5)
9:00 am - 9:45 am Wen Sun: The Role of Dataset Reset in Online Reinforcement Learning from Human Feedback
9:45 am - 10:30 pm Animesh Garg: Learning control with Differentiable Simulation
11:00 am - 11:45 am Aviral Kumar (Virtual): The "Deep" in Deep RL: Capacity Loss, and How to Mitigate it in Value-Based RL
11:45 am - 12:30 pm Max Simchowitz: Provable Guarantees for Generative Behavior Cloning
12:30 pm - 2:00 pm Lunch Break
2:00 pm - 2:45 pm Georgia Chalvatzaki (Virtual): Walking the line between model-based and learning-based methods for intelligent robotic manipulation
2:45 pm - 3:30 pm Bo Dai
Schedule (Jan 9)

Time (GMT-5)
10:30 am - 11:15 am Zhuoran Yang: Reinforcement Learning Meets Bilevel Optimization: Learning Leader-Follower Games with Sample Efficiency
11:15 am - 12:00 pm Chi Jin: Is RLHF More Difficult than Standard RL? A Theoretical Perspective
12:00 pm - 12:30 pm Contributed Talks
12:30 pm - 2:00 pm Lunch
2:00 pm - 2:45 pm Reuth Mirsky (Virtual): Goal Recognition as RL - Fantastic Goals and Where to Find Them
2:45 pm - 3:30 pm Pulkit Agrawal: Dilemmas in Reinforcement and Imitation Learning
3:30 pm - 4:15 pm Sanjiban Choudhury: To RL or not to RL


Call for Contributions

Areas of Interest
We solicit submissions related to (but not limited to) the following topics:

  • Sample Efficiency: DRL algorithms often require vast amounts of data to train. What strategies can make techniques more data-efficient, and practical for real world problems?
  • Exploration vs. Exploitation: Balancing the need to explore the environment and exploit known strategies remains a key challenge. How can we optimize this balance to yield more practical algorithms?
  • Transfer Learning: How can we leverage knowledge from one domain to enhance performance in another?
  • Stability and Convergence: With deep neural networks at their core, DRL algorithms can be prone to instability. What strategies can ensure stable learning and convergence?
  • Data and Scaling: Scaling has proven to be extremely effective in domains such as NLP and vision. How do these insights extend to deep RL algorithms?
  • Human-in-the-loop RL: RL algorithms will be deployed around people. How can these algorithms take insights from cognitive science/behavioral methods and apply them to develop provably safe/efficient DRL methods?
  • Robustness and Safety: How can we develop decision making algorithms that are provably robust and safe when deployed under uncertainty?
  • Fast Adaptation: How can we develop algorithms that are not simply provably optimal zero-shot, but rather can adapt quickly on deployment?
  • Imitation Learning: How can we develop algorithms that learn efficiently from expert demonstrations, while being able to generalize?
  • Offline Reinforcement Learning: How can we develop algorithms that provably and efficiently learn from large static datasets?
  • Real-world Application Barriers: Deploying DRL in real-world scenarios poses its own set of challenges, from safety concerns to integration issues. Our workshop will provide insights into navigating these challenges effectively.
  • Your topic can go here!

Submission Format and Instructions

We solicit papers that are 4 pages in length, preferably using the standard NeurIPS format . We are flexible though, it's alright if the paper is longer or the format is different! Please submit your papers as a PDF to this website https://forms.gle/MJuSJHJpuJ3mFevf6. Note that we do not require the papers to be anonymous and submitting them single blind is alright. Authors will present a short 3-5 minute spotlight presentation!

s
Review Guideline
  • Submissions will be evaluated based on clarity, novelty, soundness, and relevance to theme of the workshop. Both empirical and theoretical contributions are welcomed.
  • We will provide feedbacks and ideas for future improvements.
Important Dates
  • Submission deadline: Jan 1st, 2024, AoE.
  • Workshop: Jan 8th, 2024.

Papers
    TBA

Organizers

Abhishek Gupta

University of Washington

Zhaoran Wang

Northwestern University

Aldo Pacchiano

Boston University



Contact
Reach out to Abhishek Gupta (abhgupta@cs.washington.edu) or Zhaoran Wang (zhaoranwang@gmail.com) for any questions.