On Gap-dependent Bounds for Offline Reinforcement Learning
|
2023 |
Wang, Xinqi |
Don' t Pour Cereal into Coffee: Differentiable Temporal Logic for Temporal Action Segmentation
|
2023 |
Xu, Ziwei |
HUMANISE: Language-conditioned Human Motion Generation in 3D Scenes
|
2023 |
Wang, Zan |
Retaining Knowledge for Learning with Dynamic Definition
|
2023 |
Liu, Zichang |
Sequence-to-Set Generative Models
|
2023 |
Tang, Longtao |
Network change point localisation under local differential privacy
|
2023 |
Li, Mengchu |
GriddlyJS: A Web IDE for Reinforcement Learning
|
2023 |
Bamford, Christopher |
Decentralized, Communication- and Coordination-free Learning in Structured Matching Markets
|
2023 |
Maheshwari, Chinmay |
Inductive Logical Query Answering in Knowledge Graphs
|
2023 |
Galkin, Michael |
Unpacking Reward Shaping: Understanding the Benefits of Reward Engineering on Sample Complexity
|
2023 |
Gupta, Abhishek |
Retrieval-Augmented Diffusion Models
|
2023 |
Blattmann, Andreas |
STaR: Bootstrapping Reasoning With Reasoning
|
2023 |
Zelikman, Eric |
Provable Defense against Backdoor Policies in Reinforcement Learning
|
2023 |
Bharti, Shubham |
A simple but strong baseline for online continual learning: Repeated Augmented Rehearsal
|
2023 |
Zhang, Yaqian |
Compressible-composable NeRF via Rank-residual Decomposition
|
2023 |
Tang, Jiaxiang |
Data-Efficient Pipeline for Offline Reinforcement Learning with Limited Data
|
2023 |
Nie, Allen |
Hardness in Markov Decision Processes: Theory and Practice
|
2023 |
Conserva, Michelangelo |
Injecting Domain Knowledge from Empirical Interatomic Potentials to Neural Networks for Predicting Material Properties
|
2023 |
Shui, Zeren |
Keypoint-Guided Optimal Transport with Applications in Heterogeneous Domain Adaptation
|
2023 |
Gu, Xiang |
Near-Optimal Multi-Agent Learning for Safe Coverage Control
|
2023 |
Prajapat, Manish |