


default search action
Sergey Levine
Person information
- affiliation: University of California Berkeley, CA, USA
Refine list

refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
showing all ?? records
2020 – today
- 2025
- [j26]Jianlan Luo
, Charles Xu, Fangchen Liu, Liam Tan, Zipeng Lin, Jeffrey Wu, Pieter Abbeel, Sergey Levine:
FMB: A functional manipulation benchmark for generalizable robotic learning. Int. J. Robotics Res. 44(4): 592-606 (2025) - [j25]Zhongyu Li
, Xue Bin Peng, Pieter Abbeel, Sergey Levine, Glen Berseth, Koushil Sreenath:
Reinforcement learning for versatile, dynamic, and robust bipedal locomotion control. Int. J. Robotics Res. 44(5): 840-888 (2025) - [c436]Hao Bai, Yifei Zhou, Li Erran Li, Sergey Levine, Aviral Kumar:
Digi-Q: Learning VLM Q-Value Functions for Training Device-Control Agents. ICLR 2025 - [c435]Kevin Frans, Danijar Hafner, Sergey Levine, Pieter Abbeel:
One Step Diffusion via Shortcut Models. ICLR 2025 - [c434]Joey Hong, Anca D. Dragan, Sergey Levine:
Q-SFT: Q-Learning for Language Models via Supervised Fine-Tuning. ICLR 2025 - [c433]Seohong Park, Kevin Frans, Benjamin Eysenbach, Sergey Levine:
OGBench: Benchmarking Offline Goal-Conditioned RL. ICLR 2025 - [c432]Seungeun Rho, Laura M. Smith, Tianyu Li, Sergey Levine, Xue Bin Peng, Sehoon Ha:
Language Guided Skill Discovery. ICLR 2025 - [c431]Renhao Wang, Kevin Frans, Pieter Abbeel, Sergey Levine, Alexei A. Efros:
Prioritized Generative Replay. ICLR 2025 - [c430]Chenyu Wang, Masatoshi Uehara, Yichun He, Amy Wang, Avantika Lal, Tommi S. Jaakkola, Sergey Levine, Aviv Regev, Hanchen Wang, Tommaso Biancalani:
Fine-Tuning Discrete Diffusion Models via Reward Optimization with Applications to DNA and Protein Design. ICLR 2025 - [c429]Zhiyuan Zhou, Andy Peng, Qiyang Li, Sergey Levine, Aviral Kumar:
Efficient Online Reinforcement Learning Fine-Tuning Need Not Retain Offline Data. ICLR 2025 - [c428]Katie Kang, Eric Wallace, Claire J. Tomlin, Aviral Kumar, Sergey Levine:
Unfamiliar Finetuning Examples Control How Language Models Hallucinate. NAACL (Long Papers) 2025: 3600-3612 - [i511]Joshua Jones, Oier Mees, Carmelo Sferrazza, Kyle Stachowicz, Pieter Abbeel, Sergey Levine:
Beyond Sight: Finetuning Generalist Robot Policies with Heterogeneous Sensors via Language Grounding. CoRR abs/2501.04693 (2025) - [i510]Masatoshi Uehara, Yulai Zhao, Chenyu Wang, Xiner Li, Aviv Regev, Sergey Levine, Tommaso Biancalani:
Inference-Time Alignment in Diffusion Models with Reward-Guided Generation: Tutorial and Review. CoRR abs/2501.09685 (2025) - [i509]Karl Pertsch, Kyle Stachowicz, Brian Ichter, Danny Driess, Suraj Nair, Quan Vuong, Oier Mees, Chelsea Finn, Sergey Levine:
FAST: Efficient Action Tokenization for Vision-Language-Action Models. CoRR abs/2501.09747 (2025) - [i508]Tianzhe Chu, Yuexiang Zhai, Jihan Yang, Shengbang Tong, Saining Xie, Dale Schuurmans, Quoc V. Le, Sergey Levine, Yi Ma:
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training. CoRR abs/2501.17161 (2025) - [i507]Seohong Park, Qiyang Li, Sergey Levine:
Flow Q-Learning. CoRR abs/2502.02538 (2025) - [i506]Oleh Rybkin, Michal Nauman, Preston Fu, Charlie Snell, Pieter Abbeel, Sergey Levine, Aviral Kumar:
Value-Based Deep RL Scales Predictably. CoRR abs/2502.04327 (2025) - [i505]Vivek Myers, Bill Chunyuan Zheng, Anca D. Dragan, Kuan Fang, Sergey Levine:
Temporal Representation Alignment: Successor Features Enable Emergent Compositionality in Robot Instruction Following. CoRR abs/2502.05454 (2025) - [i504]Amrith Setlur, Nived Rajaraman, Sergey Levine, Aviral Kumar:
Scaling Test-Time Compute Without Verification or RL is Suboptimal. CoRR abs/2502.12118 (2025) - [i503]Masatoshi Uehara, Xingyu Su, Yulai Zhao, Xiner Li, Aviv Regev, Shuiwang Ji, Sergey Levine, Tommaso Biancalani:
Reward-Guided Iterative Refinement in Diffusion Models at Test-Time with Applications to Protein and DNA Design. CoRR abs/2502.14944 (2025) - [i502]Hao Bai, Yifei Zhou, Li Erran Li, Sergey Levine, Aviral Kumar:
Digi-Q: Learning Q-Value Functions for Training Device-Control Agents. CoRR abs/2502.15760 (2025) - [i501]Yunhai Feng, Jiaming Han, Zhuoran Yang, Xiangyu Yue, Sergey Levine, Jianlan Luo:
Reflective Planning: Vision-Language Models for Multi-Stage Long-Horizon Robotic Manipulation. CoRR abs/2502.16707 (2025) - [i500]Lucy Xiaoyang Shi, Brian Ichter, Michael Equi, Liyiming Ke, Karl Pertsch, Quan Vuong, James Tanner, Anna Walling, Haohuan Wang, Niccolo Fusai, Adrian Li-Bell, Danny Driess, Lachy Groom, Sergey Levine, Chelsea Finn:
Hi Robot: Open-Ended Instruction Following with Hierarchical Vision-Language-Action Models. CoRR abs/2502.19417 (2025) - [i499]Xiner Li, Masatoshi Uehara, Xingyu Su, Gabriele Scalia, Tommaso Biancalani, Aviv Regev, Sergey Levine, Shuiwang Ji:
Dynamic Search for Inference-Time Alignment in Diffusion Models. CoRR abs/2503.02039 (2025) - [i498]Yifei Zhou, Song Jiang, Yuandong Tian, Jason Weston, Sergey Levine, Sainbayar Sukhbaatar, Xian Li:
SWEET-RL: Training Multi-Turn LLM Agents on Collaborative Reasoning Tasks. CoRR abs/2503.15478 (2025) - [i497]Nitish Dashora, Dibya Ghosh, Sergey Levine:
ViVa: Video-Trained Value Functions for Guiding Online RL from Diverse Data. CoRR abs/2503.18210 (2025) - [i496]Zhiyuan Zhou, Pranav Atreya, You Liang Tan, Karl Pertsch, Sergey Levine:
AutoEval: Autonomous Evaluation of Generalist Robot Manipulation Policies in the Real World. CoRR abs/2503.24278 (2025) - [i495]Physical Intelligence, Kevin Black, Noah Brown, James Darpinian, Karan Dhabalia, Danny Driess, Adnan Esmail, Michael Equi, Chelsea Finn, Niccolo Fusai, Manuel Y. Galliker, Dibya Ghosh, Lachy Groom, Karol Hausman, Brian Ichter, Szymon Jakubczak, Tim Jones, Liyiming Ke, Devin LeBlanc, Sergey Levine, Adrian Li-Bell, Mohith Mothukuri, Suraj Nair, Karl Pertsch, Allen Z. Ren, Lucy Xiaoyang Shi, Laura M. Smith, Jost Tobias Springenberg, Kyle Stachowicz, James Tanner, Quan Vuong, Homer Walke, Anna Walling, Haohuan Wang, Lili Yu, Ury Zhilinsky:
π0.5: a Vision-Language-Action Model with Open-World Generalization. CoRR abs/2504.16054 (2025) - 2024
- [j24]Rafael Rafailov, Kyle Beltran Hatch, Anikait Singh, Aviral Kumar, Laura M. Smith, Ilya Kostrikov, Philippe Hansen-Estruch, Victor Kolev, Philip J. Ball, Jiajun Wu, Sergey Levine, Chelsea Finn:
D5RL: Diverse Datasets for Data-Driven Deep Reinforcement Learning. RLJ 5: 2178-2197 (2024) - [j23]Noriaki Hirose
, Dhruv Shah
, Ajay Sridhar
, Sergey Levine
:
SACSoN: Scalable Autonomous Control for Social Navigation. IEEE Robotics Autom. Lett. 9(1): 49-56 (2024) - [j22]Jianlan Luo
, Charles Xu
, Xinyang Geng
, Gilbert Feng
, Kuan Fang
, Liam Tan
, Stefan Schaal
, Sergey Levine
:
Multistage Cable Routing Through Hierarchical Imitation Learning. IEEE Trans. Robotics 40: 1476-1491 (2024) - [c427]Kuba Grudzien Kuba, Masatoshi Uehara, Sergey Levine, Pieter Abbeel:
Functional Graphical Models: Structure Enables Offline Data-Driven Optimization. AISTATS 2024: 2449-2457 - [c426]Marwa Abdulhai, Micah Carroll, Justin Svegliato, Anca D. Dragan, Sergey Levine:
Defining Deception in Decision Making. AAMAS 2024: 2111-2113 - [c425]Noriaki Hirose, Dhruv Shah, Kyle Stachowicz, Ajay Sridhar, Sergey Levine:
SELFI: Autonomous Self-Improvement with RL for Vision-Based Navigation around People. CoRL 2024: 97-116 - [c424]Ria Doshi, Homer Rich Walke, Oier Mees, Sudeep Dasari, Sergey Levine:
Scaling Cross-Embodied Learning: One Policy for Manipulation, Navigation, Locomotion and Aviation. CoRL 2024: 496-512 - [c423]Noriaki Hirose, Catherine Glossop, Ajay Sridhar, Oier Mees, Sergey Levine:
LeLaN: Learning A Language-Conditioned Navigation Policy from In-the-Wild Video. CoRL 2024: 666-688 - [c422]Kyle Stachowicz, Lydia Ignatova, Sergey Levine:
Lifelong Autonomous Improvement of Navigation Foundation Models in the Wild. CoRL 2024: 1035-1047 - [c421]Vivek Myers, Chunyuan Zheng, Oier Mees, Kuan Fang, Sergey Levine:
Policy Adaptation via Language Optimization: Decomposing Tasks for Few-Shot Imitation. CoRL 2024: 1402-1426 - [c420]Moo Jin Kim, Karl Pertsch, Siddharth Karamcheti, Ted Xiao, Ashwin Balakrishna, Suraj Nair, Rafael Rafailov, Ethan Paul Foster, Pannag R. Sanketi, Quan Vuong, Thomas Kollar, Benjamin Burchfiel, Russ Tedrake, Dorsa Sadigh, Sergey Levine, Percy Liang, Chelsea Finn:
OpenVLA: An Open-Source Vision-Language-Action Model. CoRL 2024: 2679-2713 - [c419]Michal Zawalski, William Chen, Karl Pertsch, Oier Mees, Chelsea Finn, Sergey Levine:
Robotic Control via Embodied Chain-of-Thought Reasoning. CoRL 2024: 3157-3181 - [c418]Xuanlin Li, Kyle Hsu, Jiayuan Gu, Oier Mees, Karl Pertsch, Homer Rich Walke, Chuyuan Fu, Ishikaa Lunawat, Isabel Sieh, Sean Kirmani, Sergey Levine, Jiajun Wu, Chelsea Finn, Hao Su, Quan Vuong, Ted Xiao:
Evaluating Real-World Robot Manipulation Policies in Simulation. CoRL 2024: 3705-3728 - [c417]Zhuo Xu, Hao-Tien Lewis Chiang, Zipeng Fu, Mithun George Jacob, Tingnan Zhang, Tsang-Wei Edward Lee, Wenhao Yu, Connor Schenck, David Rendleman, Dhruv Shah, Fei Xia, Jasmine Hsu, Jonathan Hoech, Pete Florence, Sean Kirmani, Sumeet Singh, Vikas Sindhwani, Carolina Parada, Chelsea Finn, Peng Xu, Sergey Levine, Jie Tan:
Mobility VLA: Multimodal Instruction Navigation with Long-Context VLMs and Topological Graphs. CoRL 2024: 3866-3887 - [c416]Zhiyuan Zhou, Pranav Atreya, Abraham Lee, Homer Rich Walke, Oier Mees, Sergey Levine:
Autonomous Improvement of Instruction Following Skills via Foundation Models. CoRL 2024: 4805-4825 - [c415]Mitsuhiko Nakamoto, Oier Mees, Aviral Kumar, Sergey Levine:
Steering Your Generalists: Improving Robotic Foundation Models via Value Guidance. CoRL 2024: 4996-5013 - [c414]Kevin Black, Michael Janner, Yilun Du, Ilya Kostrikov, Sergey Levine:
Training Diffusion Models with Reinforcement Learning. ICLR 2024 - [c413]Kevin Black, Mitsuhiko Nakamoto, Pranav Atreya, Homer Rich Walke, Chelsea Finn, Aviral Kumar, Sergey Levine:
Zero-Shot Robotic Manipulation with Pre-Trained Image-Editing Diffusion Models. ICLR 2024 - [c412]Annie S. Chen, Yoonho Lee, Amrith Setlur, Sergey Levine, Chelsea Finn:
Project and Probe: Sample-Efficient Adaptation by Interpolating Orthogonal Features. ICLR 2024 - [c411]Arnav Gudibande, Eric Wallace, Charlie Snell, Xinyang Geng, Hao Liu, Pieter Abbeel, Sergey Levine, Dawn Song:
The False Promise of Imitating Proprietary Language Models. ICLR 2024 - [c410]Joey Hong, Anca D. Dragan, Sergey Levine:
Offline RL with Observation Histories: Analyzing and Improving Sample Complexity. ICLR 2024 - [c409]Katie Kang, Amrith Setlur, Claire J. Tomlin, Sergey Levine:
Deep Neural Networks Tend To Extrapolate Predictably. ICLR 2024 - [c408]Jianlan Luo, Perry Dong, Yuexiang Zhai, Yi Ma, Sergey Levine:
RLIF: Interactive Imitation Learning as Reinforcement Learning. ICLR 2024 - [c407]Seohong Park, Oleh Rybkin, Sergey Levine:
METRA: Scalable Unsupervised RL with Metric-Aware Abstraction. ICLR 2024 - [c406]Chongyi Zheng, Benjamin Eysenbach, Homer Rich Walke, Patrick Yin, Kuan Fang, Ruslan Salakhutdinov, Sergey Levine:
Stabilizing Contrastive RL: Techniques for Robotic Goal Reaching from Offline Data. ICLR 2024 - [c405]Chengshu Li, Jacky Liang, Andy Zeng, Xinyun Chen, Karol Hausman, Dorsa Sadigh, Sergey Levine, Li Fei-Fei, Fei Xia, Brian Ichter:
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator. ICML 2024 - [c404]Jesse Farebrother, Jordi Orbay, Quan Vuong, Adrien Ali Taïga, Yevgen Chebotar, Ted Xiao, Alex Irpan, Sergey Levine, Pablo Samuel Castro, Aleksandra Faust, Aviral Kumar, Rishabh Agarwal:
Stop Regressing: Training Value Functions via Classification for Scalable Deep RL. ICML 2024 - [c403]Kevin Frans, Seohong Park, Pieter Abbeel, Sergey Levine:
Unsupervised Zero-Shot Reinforcement Learning via Functional Reward Encodings. ICML 2024 - [c402]Vivek Myers, Chongyi Zheng, Anca D. Dragan, Sergey Levine, Benjamin Eysenbach:
Learning Temporal Distances: Contrastive Successor Features Can Provide a Metric Structure for Decision-Making. ICML 2024 - [c401]Soroush Nasiriany, Fei Xia, Wenhao Yu, Ted Xiao, Jacky Liang, Ishita Dasgupta, Annie Xie, Danny Driess, Ayzaan Wahid, Zhuo Xu, Quan Vuong, Tingnan Zhang, Tsang-Wei Edward Lee, Kuang-Huei Lee, Peng Xu, Sean Kirmani, Yuke Zhu, Andy Zeng, Karol Hausman, Nicolas Heess, Chelsea Finn, Sergey Levine, Brian Ichter:
PIVOT: Iterative Visual Prompting Elicits Actionable Knowledge for VLMs. ICML 2024 - [c400]Seohong Park, Tobias Kreiman, Sergey Levine:
Foundation Policies with Hilbert Representations. ICML 2024 - [c399]Amrith Setlur, Saurabh Garg, Virginia Smith, Sergey Levine:
Prompting is a Double-Edged Sword: Improving Worst-Group Robustness of Foundation Models. ICML 2024 - [c398]Masatoshi Uehara, Yulai Zhao, Kevin Black, Ehsan Hajiramezanali, Gabriele Scalia, Nathaniel Lee Diamant, Alex M. Tseng, Sergey Levine, Tommaso Biancalani:
Feedback Efficient Online Fine-Tuning of Diffusion Models. ICML 2024 - [c397]Annie Xie, Logan M. Bhamidipaty, Evan Zheran Liu, Joey Hong, Sergey Levine, Chelsea Finn:
Learning to Explore in POMDPs with Informational Rewards. ICML 2024 - [c396]Yifei Zhou, Andrea Zanette, Jiayi Pan, Sergey Levine, Aviral Kumar:
ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL. ICML 2024 - [c395]Ajay Sridhar, Dhruv Shah, Catherine Glossop, Sergey Levine:
NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration. ICRA 2024: 63-70 - [c394]Abby O'Neill, Abdul Rehman, Abhiram Maddukuri, Abhishek Gupta, Abhishek Padalkar, Abraham Lee, Acorn Pooley, Agrim Gupta, Ajay Mandlekar, Ajinkya Jain, Albert Tung, Alex Bewley, Alexander Herzog, Alex Irpan, Alexander Khazatsky, Anant Rai, Anchit Gupta, Andrew E. Wang, Anikait Singh, Animesh Garg, Aniruddha Kembhavi, Annie Xie, Anthony Brohan, Antonin Raffin, Archit Sharma, Arefeh Yavary, Arhan Jain, Ashwin Balakrishna, Ayzaan Wahid, Ben Burgess-Limerick, Beomjoon Kim, Bernhard Schölkopf, Blake Wulfe, Brian Ichter, Cewu Lu, Charles Xu, Charlotte Le, Chelsea Finn, Chen Wang, Chenfeng Xu, Cheng Chi, Chenguang Huang, Christine Chan, Christopher Agia, Chuer Pan, Chuyuan Fu, Coline Devin, Danfei Xu, Daniel Morton, Danny Driess, Daphne Chen, Deepak Pathak, Dhruv Shah, Dieter Büchler, Dinesh Jayaraman, Dmitry Kalashnikov, Dorsa Sadigh, Edward Johns, Ethan Paul Foster, Fangchen Liu, Federico Ceola, Fei Xia, Feiyu Zhao, Freek Stulp, Gaoyue Zhou, Gaurav S. Sukhatme, Gautam Salhotra, Ge Yan, Gilbert Feng, Giulio Schiavi, Glen Berseth, Gregory Kahn, Guanzhi Wang, Hao Su, Haoshu Fang, Haochen Shi, Henghui Bao, Heni Ben Amor, Henrik I. Christensen, Hiroki Furuta, Homer Walke, Hongjie Fang, Huy Ha, Igor Mordatch, Ilija Radosavovic, Isabel Leal, Jacky Liang, Jad Abou-Chakra, Jaehyung Kim, Jaimyn Drake, Jan Peters, Jan Schneider, Jasmine Hsu, Jeannette Bohg, Jeffrey Bingham, Jeffrey Wu, Jensen Gao, Jiaheng Hu, Jiajun Wu, Jialin Wu, Jiankai Sun, Jianlan Luo, Jiayuan Gu, Jie Tan, Jihoon Oh, Jimmy Wu, Jingpei Lu, Jingyun Yang, Jitendra Malik, João Silvério, Joey Hejna, Jonathan Booher, Jonathan Tompson, Jonathan Yang, Jordi Salvador, Joseph J. Lim, Junhyek Han, Kaiyuan Wang, Kanishka Rao, Karl Pertsch, Karol Hausman, Keegan Go, Keerthana Gopalakrishnan, Ken Goldberg, Kendra Byrne, Kenneth Oslund, Kento Kawaharazuka, Kevin Black, Kevin Lin, Kevin Zhang, Kiana Ehsani, Kiran Lekkala, Kirsty Ellis, Krishan Rana, Krishnan Srinivasan, Kuan Fang, Kunal Pratap Singh, Kuo-Hao Zeng, Kyle Hatch, Kyle Hsu, Laurent Itti, Lawrence Yunliang Chen, Lerrel Pinto, Li Fei-Fei, Liam Tan, Linxi Jim Fan, Lionel Ott, Lisa Lee, Luca Weihs, Magnum Chen, Marion Lepert, Marius Memmel, Masayoshi Tomizuka, Masha Itkina, Mateo Guaman Castro, Max Spero, Maximilian Du, Michael Ahn, Michael C. Yip, Mingtong Zhang, Mingyu Ding, Minho Heo, Mohan Kumar Srirama, Mohit Sharma, Moo Jin Kim, Naoaki Kanazawa, Nicklas Hansen, Nicolas Heess, Nikhil J. Joshi, Niko Sünderhauf, Ning Liu, Norman Di Palo, Nur Muhammad (Mahi) Shafiullah, Oier Mees, Oliver Kroemer, Osbert Bastani, Pannag R. Sanketi, Patrick Tree Miller, Patrick Yin, Paul Wohlhart, Peng Xu, Peter David Fagan, Peter Mitrano, Pierre Sermanet, Pieter Abbeel, Priya Sundaresan, Qiuyu Chen, Quan Vuong, Rafael Rafailov, Ran Tian, Ria Doshi, Roberto Martín-Martín, Rohan Baijal, Rosario Scalise, Rose Hendrix, Roy Lin, Runjia Qian, Ruohan Zhang, Russell Mendonca, Rutav Shah, Ryan Hoque, Ryan Julian, Samuel Bustamante, Sean Kirmani, Sergey Levine, Shan Lin, Sherry Moore, Shikhar Bahl, Shivin Dass, Shubham D. Sonawani, Shuran Song, Sichun Xu, Siddhant Haldar, Siddharth Karamcheti, Simeon Adebola, Simon Guist, Soroush Nasiriany, Stefan Schaal, Stefan Welker, Stephen Tian, Subramanian Ramamoorthy, Sudeep Dasari, Suneel Belkhale, Sungjae Park, Suraj Nair, Suvir Mirchandani, Takayuki Osa, Tanmay Gupta, Tatsuya Harada, Tatsuya Matsushima, Ted Xiao, Thomas Kollar, Tianhe Yu, Tianli Ding, Todor Davchev, Tony Z. Zhao, Travis Armstrong, Trevor Darrell, Trinity Chung, Vidhi Jain, Vincent Vanhoucke, Wei Zhan, Wenxuan Zhou, Wolfram Burgard, Xi Chen, Xiaolong Wang, Xinghao Zhu, Xinyang Geng, Xiyuan Liu, Liangwei Xu, Xuanlin Li, Yao Lu, Yecheng Jason Ma, Yejin Kim, Yevgen Chebotar, Yifan Zhou, Yifeng Zhu, Yilin Wu, Ying Xu, Yixuan Wang, Yonatan Bisk, Yoonyoung Cho, Youngwoon Lee, Yuchen Cui, Yue Cao, Yueh-Hua Wu, Yujin Tang, Yuke Zhu, Yunchu Zhang, Yunfan Jiang, Yunshuang Li, Yunzhu Li, Yusuke Iwasawa, Yutaka Matsuo, Zehan Ma, Zhuo Xu, Zichen Jeff Cui, Zichen Zhang, Zipeng Lin:
Open X-Embodiment: Robotic Learning Datasets and RT-X Models : Open X-Embodiment Collaboration. ICRA 2024: 6892-6903 - [c393]Laura M. Smith, Yunhao Cao, Sergey Levine:
Grow Your Limits: Continuous Improvement with Real-World RL for Robotic Locomotion. ICRA 2024: 10829-10836 - [c392]Jianlan Luo, Zheyuan Hu
, Charles Xu, You Liang Tan, Jacob Berg, Archit Sharma, Stefan Schaal, Chelsea Finn, Abhishek Gupta, Sergey Levine:
SERL: A Software Suite for Sample-Efficient Robotic Reinforcement Learning. ICRA 2024: 16961-16969 - [c391]Chethan Bhateja, Derek Guo, Dibya Ghosh, Anikait Singh, Manan Tomar, Quan Vuong, Yevgen Chebotar, Sergey Levine, Aviral Kumar:
Robotic Offline RL from Internet Videos via Value-Function Learning. ICRA 2024: 16977-16984 - [c390]Xiaoyu Huang, Qiayuan Liao, Yiming Ni
, Zhongyu Li, Laura M. Smith, Sergey Levine, Xue Bin Peng, Koushil Sreenath:
HiLMa-Res: A General Hierarchical Framework via Residual RL for Combining Quadrupedal Locomotion and Manipulation. IROS 2024: 9050-9057 - [c389]Hao Bai, Yifei Zhou, Jiayi Pan, Mert Cemri, Alane Suhr, Sergey Levine, Aviral Kumar:
DigiRL: Training In-The-Wild Device-Control Agents with Autonomous Reinforcement Learning. NeurIPS 2024 - [c388]Benjamin Eysenbach, Vivek Myers, Ruslan Salakhutdinov, Sergey Levine:
Inference via Interpolation: Contrastive Representations Provably Enable Planning and Inference. NeurIPS 2024 - [c387]Vivek Myers, Evan Ellis, Sergey Levine, Benjamin Eysenbach, Anca D. Dragan:
Learning to Assist Humans without Inferring Rewards. NeurIPS 2024 - [c386]Seohong Park, Kevin Frans, Sergey Levine, Aviral Kumar:
Is Value Learning Really the Main Bottleneck in Offline RL? NeurIPS 2024 - [c385]Aniketh Janardhan Reddy, Xinyang Geng, Michael Herschl, Sathvik Kolli, Aviral Kumar, Patrick Hsu, Sergey Levine, Nilah Ioannidis:
Designing Cell-Type-Specific Promoter Sequences Using Conservative Model-Based Optimization. NeurIPS 2024 - [c384]Masatoshi Uehara, Yulai Zhao, Ehsan Hajiramezanali, Gabriele Scalia, Gökcen Eraslan, Avantika Lal, Sergey Levine, Tommaso Biancalani:
Bridging Model-Based Optimization and Generative Modeling via Conservative Fine-Tuning of Diffusion Models. NeurIPS 2024 - [c383]Simon Zhai, Hao Bai, Zipeng Lin, Jiayi Pan, Peter Tong, Yifei Zhou, Alane Suhr, Saining Xie, Yann LeCun, Yi Ma, Sergey Levine:
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning. NeurIPS 2024 - [c382]Kuan Fang, Fangchen Liu, Pieter Abbeel, Sergey Levine:
MOKA: Open-World Robotic Manipulation through Mark-Based Visual Prompting. Robotics: Science and Systems 2024 - [c381]Dibya Ghosh, Homer Rich Walke, Karl Pertsch, Kevin Black, Oier Mees, Sudeep Dasari, Joey Hejna, Tobias Kreiman, Charles Xu, Jianlan Luo, You Liang Tan, Lawrence Yunliang Chen, Quan Vuong, Ted Xiao, Pannag R. Sanketi, Dorsa Sadigh, Chelsea Finn, Sergey Levine:
Octo: An Open-Source Generalist Robot Policy. Robotics: Science and Systems 2024 - [c380]Alexander Khazatsky, Karl Pertsch, Suraj Nair, Ashwin Balakrishna, Sudeep Dasari, Siddharth Karamcheti, Soroush Nasiriany, Mohan Kumar Srirama, Lawrence Yunliang Chen, Kirsty Ellis, Peter David Fagan, Joey Hejna, Masha Itkina, Marion Lepert, Yecheng Jason Ma, Patrick Tree Miller, Jimmy Wu, Suneel Belkhale, Shivin Dass, Huy Ha, Arhan Jain, Abraham Lee, Youngwoon Lee, Marius Memmel, Sungjae Park, Ilija Radosavovic, Kaiyuan Wang, Albert Zhan, Kevin Black, Cheng Chi, Kyle Beltran Hatch, Shan Lin, Jingpei Lu, Jean Mercat, Abdul Rehman, Pannag R. Sanketi, Archit Sharma, Cody Simpson, Quan Vuong, Homer Rich Walke, Blake Wulfe, Ted Xiao, Jonathan Heewon Yang, Arefeh Yavary, Tony Z. Zhao, Christopher Agia, Rohan Baijal, Mateo Guaman Castro, Daphne Chen, Qiuyu Chen, Trinity Chung, Jaimyn Drake, Ethan Paul Foster, Jensen Gao, David Antonio Herrera, Minho Heo, Kyle Hsu, Jiaheng Hu, Donovon Jackson, Charlotte Le, Yunshuang Li, Roy Lin, Zehan Ma, Abhiram Maddukuri, Suvir Mirchandani, Daniel Morton, Tony Nguyen, Abigail O'Neill, Rosario Scalise, Derick Seale, Victor Son, Stephen Tian, Emi Tran, Andrew E. Wang, Yilin Wu, Annie Xie, Jingyun Yang, Patrick Yin, Yunchu Zhang, Osbert Bastani, Glen Berseth, Jeannette Bohg
, Ken Goldberg, Abhinav Gupta, Abhishek Gupta, Dinesh Jayaraman, Joseph J. Lim, Jitendra Malik, Roberto Martín-Martín, Subramanian Ramamoorthy, Dorsa Sadigh, Shuran Song, Jiajun Wu, Michael C. Yip, Yuke Zhu, Thomas Kollar, Sergey Levine, Chelsea Finn:
DROID: A Large-Scale In-The-Wild Robot Manipulation Dataset. Robotics: Science and Systems 2024 - [c379]Lucy Xiaoyang Shi, Zheyuan Hu, Tony Z. Zhao, Archit Sharma, Karl Pertsch, Jianlan Luo, Sergey Levine, Chelsea Finn:
Yell At Your Robot: Improving On-the-Fly from Language Corrections. Robotics: Science and Systems 2024 - [c378]Kyle Stachowicz, Sergey Levine:
RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes. Robotics: Science and Systems 2024 - [c377]Jonathan Heewon Yang, Catherine Glossop, Arjun Bhorkar, Dhruv Shah, Quan Vuong, Chelsea Finn, Dorsa Sadigh, Sergey Levine:
Pushing the Limits of Cross-Embodiment Learning for Manipulation and Navigation. Robotics: Science and Systems 2024 - [i494]Jakub Grudzien Kuba, Masatoshi Uehara, Pieter Abbeel, Sergey Levine:
Functional Graphical Models: Structure Enables Offline Data-Driven Optimization. CoRR abs/2401.05442 (2024) - [i493]Jianlan Luo, Charles Xu, Fangchen Liu, Liam Tan, Zipeng Lin, Jeffrey Wu, Pieter Abbeel, Sergey Levine:
FMB: a Functional Manipulation Benchmark for Generalizable Robotic Learning. CoRR abs/2401.08553 (2024) - [i492]Michael Ahn, Debidatta Dwibedi, Chelsea Finn, Montse Gonzalez Arenas, Keerthana Gopalakrishnan, Karol Hausman, Brian Ichter, Alex Irpan, Nikhil J. Joshi, Ryan Julian, Sean Kirmani, Isabel Leal, Tsang-Wei Edward Lee, Sergey Levine, Yao Lu, Sharath Maddineni, Kanishka Rao, Dorsa Sadigh, Pannag Sanketi, Pierre Sermanet, Quan Vuong, Stefan Welker, Fei Xia, Ted Xiao, Peng Xu, Steve Xu, Zhuo Xu:
AutoRT: Embodied Foundation Models for Large Scale Orchestration of Robotic Agents. CoRR abs/2401.12963 (2024) - [i491]Jianlan Luo, Zheyuan Hu, Charles Xu, You Liang Tan, Jacob Berg, Archit Sharma, Stefan Schaal, Chelsea Finn, Abhishek Gupta, Sergey Levine:
SERL: A Software Suite for Sample-Efficient Robotic Reinforcement Learning. CoRR abs/2401.16013 (2024) - [i490]Zhongyu Li, Xue Bin Peng, Pieter Abbeel, Sergey Levine, Glen Berseth, Koushil Sreenath:
Reinforcement Learning for Versatile, Dynamic, and Robust Bipedal Locomotion Control. CoRR abs/2401.16889 (2024) - [i489]William Chen, Oier Mees, Aviral Kumar, Sergey Levine:
Vision-Language Models Provide Promptable Representations for Reinforcement Learning. CoRR abs/2402.02651 (2024) - [i488]Soroush Nasiriany, Fei Xia, Wenhao Yu, Ted Xiao, Jacky Liang, Ishita Dasgupta, Annie Xie, Danny Driess, Ayzaan Wahid, Zhuo Xu, Quan Vuong, Tingnan Zhang, Tsang-Wei Edward Lee, Kuang-Huei Lee, Peng Xu, Sean Kirmani, Yuke Zhu, Andy Zeng, Karol Hausman, Nicolas Heess, Chelsea Finn, Sergey Levine, Brian Ichter:
PIVOT: Iterative Visual Prompting Elicits Actionable Knowledge for VLMs. CoRR abs/2402.07872 (2024) - [i487]Masatoshi Uehara, Yulai Zhao, Kevin Black, Ehsan Hajiramezanali, Gabriele Scalia, Nathaniel Lee Diamant, Alex M. Tseng, Tommaso Biancalani, Sergey Levine:
Fine-Tuning of Continuous-Time Diffusion Models as Entropy-Regularized Control. CoRR abs/2402.15194 (2024) - [i486]Seohong Park, Tobias Kreiman, Sergey Levine:
Foundation Policies with Hilbert Representations. CoRR abs/2402.15567 (2024) - [i485]Masatoshi Uehara, Yulai Zhao, Kevin Black, Ehsan Hajiramezanali, Gabriele Scalia, Nathaniel Lee Diamant, Alex M. Tseng, Sergey Levine, Tommaso Biancalani:
Feedback Efficient Online Fine-Tuning of Diffusion Models. CoRR abs/2402.16359 (2024) - [i484]Kevin Frans, Seohong Park, Pieter Abbeel, Sergey Levine:
Unsupervised Zero-Shot Reinforcement Learning via Functional Reward Encodings. CoRR abs/2402.17135 (2024) - [i483]Jonathan Yang, Catherine Glossop, Arjun Bhorkar, Dhruv Shah, Quan Vuong, Chelsea Finn, Dorsa Sadigh, Sergey Levine:
Pushing the Limits of Cross-Embodiment Learning for Manipulation and Navigation. CoRR abs/2402.19432 (2024) - [i482]Yifei Zhou, Andrea Zanette, Jiayi Pan, Sergey Levine, Aviral Kumar:
ArCHer: Training Language Model Agents via Hierarchical Multi-Turn RL. CoRR abs/2402.19446 (2024) - [i481]Noriaki Hirose, Dhruv Shah, Kyle Stachowicz, Ajay Sridhar, Sergey Levine:
SELFI: Autonomous Self-Improvement with Reinforcement Learning for Social Navigation. CoRR abs/2403.00991 (2024) - [i480]Fangchen Liu, Kuan Fang, Pieter Abbeel, Sergey Levine:
MOKA: Open-Vocabulary Robotic Manipulation through Mark-Based Visual Prompting. CoRR abs/2403.03174 (2024) - [i479]Jesse Farebrother, Jordi Orbay, Quan Vuong, Adrien Ali Taïga, Yevgen Chebotar, Ted Xiao, Alex Irpan, Sergey Levine, Pablo Samuel Castro, Aleksandra Faust, Aviral Kumar, Rishabh Agarwal:
Stop Regressing: Training Value Functions via Classification for Scalable Deep RL. CoRR abs/2403.03950 (2024) - [i478]Benjamin Eysenbach, Vivek Myers, Ruslan Salakhutdinov, Sergey Levine:
Inference via Interpolation: Contrastive Representations Provably Enable Planning and Inference. CoRR abs/2403.04082 (2024) - [i477]Katie Kang, Eric Wallace, Claire J. Tomlin, Aviral Kumar, Sergey Levine:
Unfamiliar Finetuning Examples Control How Language Models Hallucinate. CoRR abs/2403.05612 (2024) - [i476]Lucy Xiaoyang Shi, Zheyuan Hu, Tony Z. Zhao, Archit Sharma, Karl Pertsch, Jianlan Luo, Sergey Levine, Chelsea Finn:
Yell At Your Robot: Improving On-the-Fly from Language Corrections. CoRR abs/2403.12910 (2024) - [i475]Jiayi Pan, Yichi Zhang
, Nicholas Tomlin, Yifei Zhou, Sergey Levine, Alane Suhr:
Autonomous Evaluation and Refinement of Digital Agents. CoRR abs/2404.06474 (2024) - [i474]Toru Lin, Yu Zhang, Qiyang Li, Haozhi Qi, Brent Yi, Sergey Levine, Jitendra Malik:
Learning Visuotactile Skills with Two Multifingered Hands. CoRR abs/2404.16823 (2024) - [i473]Kyle Stachowicz, Sergey Levine:
RACER: Epistemic Risk-Sensitive RL Enables Fast Driving with Fewer Crashes. CoRR abs/2405.04714 (2024) - [i472]Xuanlin Li, Kyle Hsu, Jiayuan Gu, Karl Pertsch, Oier Mees, Homer Rich Walke, Chuyuan Fu, Ishikaa Lunawat, Isabel Sieh, Sean Kirmani, Sergey Levine, Jiajun Wu, Chelsea Finn, Hao Su, Quan Vuong, Ted Xiao:
Evaluating Real-World Robot Manipulation Policies in Simulation. CoRR abs/2405.05941 (2024) - [i471]Yuexiang Zhai, Hao Bai, Zipeng Lin, Jiayi Pan, Shengbang Tong, Yifei Zhou, Alane Suhr, Saining Xie, Yann LeCun, Yi Ma, Sergey Levine:
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning. CoRR abs/2405.10292 (2024) - [i470]Octo Model Team, Dibya Ghosh, Homer Walke, Karl Pertsch, Kevin Black, Oier Mees, Sudeep Dasari, Joey Hejna, Tobias Kreiman, Charles Xu, Jianlan Luo, You Liang Tan, Lawrence Yunliang Chen, Pannag Sanketi, Quan Vuong, Ted Xiao, Dorsa Sadigh, Chelsea Finn, Sergey Levine:
Octo: An Open-Source Generalist Robot Policy. CoRR abs/2405.12213 (2024) - [i469]Masatoshi Uehara, Yulai Zhao, Ehsan Hajiramezanali, Gabriele Scalia, Gökcen Eraslan, Avantika Lal, Sergey Levine, Tommaso Biancalani:
Bridging Model-Based Optimization and Generative Modeling via Conservative Fine-Tuning of Diffusion Models. CoRR abs/2405.19673 (2024) - [i468]Yutaka Shimizu, Joey Hong, Sergey Levine, Masayoshi Tomizuka:
Strategically Conservative Q-Learning. CoRR abs/2406.04534 (2024) - [i467]Seungeun Rho, Laura M. Smith, Tianyu Li, Sergey Levine, Xue Bin Peng, Sehoon Ha:
Language Guided Skill Discovery. CoRR abs/2406.06615 (2024) - [i466]Moo Jin Kim, Karl Pertsch, Siddharth Karamcheti, Ted Xiao, Ashwin Balakrishna, Suraj Nair, Rafael Rafailov, Ethan Paul Foster, Grace Lam, Pannag Sanketi, Quan Vuong, Thomas Kollar, Benjamin Burchfiel, Russ Tedrake, Dorsa Sadigh, Sergey Levine, Percy Liang, Chelsea Finn:
OpenVLA: An Open-Source Vision-Language-Action Model. CoRR abs/2406.09246 (2024) - [i465]Seohong Park, Kevin Frans, Sergey Levine, Aviral Kumar:
Is Value Learning Really the Main Bottleneck in Offline RL? CoRR abs/2406.09329 (2024) - [i464]Hao Bai, Yifei Zhou, Mert Cemri, Jiayi Pan, Alane Suhr, Sergey Levine, Aviral Kumar:
DigiRL: Training In-The-Wild Device-Control Agents with Autonomous Reinforcement Learning. CoRR abs/2406.11896 (2024) - [i463]Yulai Zhao, Masatoshi Uehara, Gabriele Scalia, Tommaso Biancalani, Sergey Levine, Ehsan Hajiramezanali:
Adding Conditional Control to Diffusion Models with Reinforcement Learning. CoRR abs/2406.12120 (2024) - [i462]Vivek Myers, Chongyi Zheng, Anca D. Dragan, Sergey Levine, Benjamin Eysenbach:
Learning Temporal Distances: Contrastive Successor Features Can Provide a Metric Structure for Decision-Making. CoRR abs/2406.17098 (2024) - [i461]Annie S. Chen, Alec M. Lessing, Andy Tang, Govind Chada, Laura M. Smith, Sergey Levine, Chelsea Finn:
Commonsense Reasoning for Legged Robot Adaptation with Vision-Language Models. CoRR abs/2407.02666 (2024) - [i460]Xiaoyu Huang, Qiayuan Liao, Yiming Ni, Zhongyu Li, Laura M. Smith, Sergey Levine, Xue Bin Peng, Koushil Sreenath:
HiLMa-Res: A General Hierarchical Framework via Residual RL for Combining Quadrupedal Locomotion and Manipulation. CoRR abs/2407.06584 (2024) - [i459]Hao-Tien Lewis Chiang, Zhuo Xu, Zipeng Fu, Mithun George Jacob, Tingnan Zhang, Tsang-Wei Edward Lee, Wenhao Yu, Connor Schenck, David Rendleman, Dhruv Shah, Fei Xia, Jasmine Hsu, Jonathan Hoech, Pete Florence, Sean Kirmani, Sumeet Singh, Vikas Sindhwani, Carolina Parada, Chelsea Finn, Peng Xu, Sergey Levine, Jie Tan:
Mobility VLA: Multimodal Instruction Navigation with Long-Context VLMs and Topological Graphs. CoRR abs/2407.07775 (2024) - [i458]Michal Zawalski, William Chen, Karl Pertsch, Oier Mees, Chelsea Finn, Sergey Levine:
Robotic Control via Embodied Chain-of-Thought Reasoning. CoRR abs/2407.08693 (2024) - [i457]Manan Tomar, Philippe Hansen-Estruch, Philip Bachman, Alex Lamb, John Langford, Matthew E. Taylor, Sergey Levine:
Video Occupancy Models. CoRR abs/2407.09533 (2024) - [i456]Masatoshi Uehara, Yulai Zhao, Tommaso Biancalani, Sergey Levine:
Understanding Reinforcement Learning-Based Fine-Tuning of Diffusion Models: A Tutorial and Review. CoRR abs/2407.13734 (2024) - [i455]Zhiyuan Zhou, Pranav Atreya, Abraham Lee, Homer Walke, Oier Mees, Sergey Levine:
Autonomous Improvement of Instruction Following Skills via Foundation Models. CoRR abs/2407.20635 (2024) - [i454]Xiner Li, Yulai Zhao, Chenyu Wang, Gabriele Scalia, Gökcen Eraslan, Surag Nair, Tommaso Biancalani, Aviv Regev, Sergey Levine, Masatoshi Uehara:
Derivative-Free Guidance in Continuous and Discrete Diffusion Models with Soft Value-Based Decoding. CoRR abs/2408.08252 (2024) - [i453]Rafael Rafailov, Kyle Hatch, Anikait Singh, Laura M. Smith, Aviral Kumar, Ilya Kostrikov, Philippe Hansen-Estruch, Victor Kolev, Philip J. Ball, Jiajun Wu, Chelsea Finn, Sergey Levine:
D5RL: Diverse Datasets for Data-Driven Deep Reinforcement Learning. CoRR abs/2408.08441 (2024) - [i452]Ria Doshi, Homer Walke, Oier Mees, Sudeep Dasari, Sergey Levine:
Scaling Cross-Embodied Learning: One Policy for Manipulation, Navigation, Locomotion and Aviation. CoRR abs/2408.11812 (2024) - [i451]Junsu Kim, Seohong Park, Sergey Levine:
Unsupervised-to-Online Reinforcement Learning. CoRR abs/2408.14785 (2024) - [i450]Vivek Myers, Bill Chunyuan Zheng, Oier Mees, Sergey Levine, Kuan Fang:
Policy Adaptation via Language Optimization: Decomposing Tasks for Few-Shot Imitation. CoRR abs/2408.16228 (2024) - [i449]Grace Tang, Swetha Rajkumar, Yifei Zhou, Homer Rich Walke, Sergey Levine, Kuan Fang:
KALIE: Fine-Tuning Vision-Language Models for Open-World Manipulation without Robot Data. CoRR abs/2409.14066 (2024) - [i448]Noriaki Hirose, Catherine Glossop, Ajay Sridhar, Dhruv Shah, Oier Mees, Sergey Levine:
LeLaN: Learning A Language-Conditioned Navigation Policy from In-the-Wild Videos. CoRR abs/2410.03603 (2024) - [i447]Sudeep Dasari, Oier Mees, Sebastian Zhao, Mohan Kumar Srirama, Sergey Levine:
The Ingredients for Robotic Diffusion Transformers. CoRR abs/2410.10088 (2024) - [i446]Hongbo Zhang, Zhongyu Li, Xuanqi Zeng, Laura M. Smith, Kyle Stachowicz, Dhruv Shah, Linzhu Yue, Zhitao Song, Weipeng Xia, Sergey Levine, Koushil Sreenath, Yun-hui Liu:
Traversability-Aware Legged Navigation by Learning from Real-World Visual Data. CoRR abs/2410.10621 (2024) - [i445]Kevin Frans, Danijar Hafner, Sergey Levine, Pieter Abbeel:
One Step Diffusion via Shortcut Models. CoRR abs/2410.12557 (2024) - [i444]Jakub Grudzien Kuba, Pieter Abbeel, Sergey Levine:
Cliqueformer: Model-Based Optimization with Structured Transformers. CoRR abs/2410.13106 (2024) - [i443]Chenyu Wang, Masatoshi Uehara, Yichun He, Amy Wang, Tommaso Biancalani, Avantika Lal, Tommi S. Jaakkola, Sergey Levine, Hanchen Wang, Aviv Regev:
Fine-Tuning Discrete Diffusion Models via Reward Optimization with Applications to DNA and Protein Design. CoRR abs/2410.13643 (2024) - [i442]Mitsuhiko Nakamoto, Oier Mees, Aviral Kumar, Sergey Levine:
Steering Your Generalists: Improving Robotic Foundation Models via Value Guidance. CoRR abs/2410.13816 (2024) - [i441]Max Wilcoxson, Qiyang Li, Kevin Frans, Sergey Levine:
Leveraging Skills from Unlabeled Prior Data for Efficient Online Exploration. CoRR abs/2410.18076 (2024) - [i440]Renhao Wang, Kevin Frans, Pieter Abbeel, Sergey Levine, Alexei A. Efros:
Prioritized Generative Replay. CoRR abs/2410.18082 (2024) - [i439]Kyle Beltran Hatch, Ashwin Balakrishna, Oier Mees, Suraj Nair, Seohong Park, Blake Wulfe, Masha Itkina, Benjamin Eysenbach, Sergey Levine, Thomas Kollar, Benjamin Burchfiel:
GHIL-Glue: Hierarchical Control with Filtered Subgoal Images. CoRR abs/2410.20018 (2024) - [i438]Seohong Park, Kevin Frans, Benjamin Eysenbach, Sergey Levine:
OGBench: Benchmarking Offline Goal-Conditioned RL. CoRR abs/2410.20092 (2024) - [i437]Jianlan Luo, Charles Xu, Jeffrey Wu, Sergey Levine:
Precise and Dexterous Robotic Manipulation via Human-in-the-Loop Reinforcement Learning. CoRR abs/2410.21845 (2024) - [i436]Kevin Black, Noah Brown, Danny Driess, Adnan Esmail, Michael Equi, Chelsea Finn, Niccolo Fusai, Lachy Groom, Karol Hausman, Brian Ichter, Szymon Jakubczak, Tim Jones, Liyiming Ke
, Sergey Levine, Adrian Li-Bell, Mohith Mothukuri, Suraj Nair, Karl Pertsch, Lucy Xiaoyang Shi, James Tanner, Quan Vuong, Anna Walling, Haohuan Wang, Ury Zhilinsky:
π0: A Vision-Language-Action Flow Model for General Robot Control. CoRR abs/2410.24164 (2024) - [i435]Vivek Myers, Evan Ellis, Sergey Levine, Benjamin Eysenbach, Anca D. Dragan:
Learning to Assist Humans without Inferring Rewards. CoRR abs/2411.02623 (2024) - [i434]Joey Hong, Anca D. Dragan, Sergey Levine:
Q-SFT: Q-Learning for Language Models via Supervised Fine-Tuning. CoRR abs/2411.05193 (2024) - [i433]Joey Hong, Jessica Lin, Anca D. Dragan, Sergey Levine:
Interactive Dialogue Agents via Reinforcement Learning on Hindsight Regenerations. CoRR abs/2411.05194 (2024) - [i432]Katie Kang, Amrith Setlur, Dibya Ghosh, Jacob Steinhardt, Claire J. Tomlin, Sergey Levine, Aviral Kumar:
What Do Learning Dynamics Reveal About Generalization in LLM Reasoning? CoRR abs/2411.07681 (2024) - [i431]Charlie Snell, Eric Wallace, Dan Klein, Sergey Levine:
Predicting Emergent Capabilities by Finetuning. CoRR abs/2411.16035 (2024) - [i430]Zhiyuan Zhou, Andy Peng, Qiyang Li, Sergey Levine, Aviral Kumar:
Efficient Online Reinforcement Learning Fine-Tuning Need Not Retain Offline Data. CoRR abs/2412.07762 (2024) - [i429]Charles Xu, Qiyang Li, Jianlan Luo, Sergey Levine:
RLDG: Robotic Generalist Policy Distillation via Reinforcement Learning. CoRR abs/2412.09858 (2024) - [i428]Yifei Zhou, Qianlan Yang, Kaixiang Lin, Min Bai, Xiong Zhou, Yu-Xiong Wang, Sergey Levine, Li Erran Li:
Proposer-Agent-Evaluator(PAE): Autonomous Skill Discovery For Foundation Model Internet Agents. CoRR abs/2412.13194 (2024) - 2023
- [j21]Shagun Sodhani, Sergey Levine, Amy Zhang:
Improving Generalization with Approximate Factored Value Functions. Trans. Mach. Learn. Res. 2023 (2023) - [c376]Dhruv Shah, Ajay Sridhar, Nitish Dashora, Kyle Stachowicz, Kevin Black, Noriaki Hirose, Sergey Levine:
ViNT: A Foundation Model for Visual Navigation. CoRL 2023: 711-733 - [c375]Jianlan Luo, Perry Dong, Jeffrey Wu, Aviral Kumar, Xinyang Geng, Sergey Levine:
Action-Quantized Offline Reinforcement Learning for Robotic Skill Learning. CoRL 2023: 1348-1361 - [c374]Homer Rich Walke, Kevin Black, Tony Z. Zhao, Quan Vuong, Chongyi Zheng, Philippe Hansen-Estruch, Andre Wang He, Vivek Myers, Moo Jin Kim, Max Du, Abraham Lee, Kuan Fang, Chelsea Finn, Sergey Levine:
BridgeData V2: A Dataset for Robot Learning at Scale. CoRL 2023: 1723-1736 - [c373]Zheyuan Hu
, Aaron Rovinsky, Jianlan Luo, Vikash Kumar, Abhishek Gupta, Sergey Levine:
REBOOT: Reuse Data for Bootstrapping Efficient Real-World Dexterous Manipulation. CoRL 2023: 1930-1949 - [c372]Brianna Zitkovich, Tianhe Yu, Sichun Xu, Peng Xu, Ted Xiao, Fei Xia, Jialin Wu, Paul Wohlhart, Stefan Welker, Ayzaan Wahid, Quan Vuong, Vincent Vanhoucke, Huong T. Tran, Radu Soricut, Anikait Singh, Jaspiar Singh, Pierre Sermanet, Pannag R. Sanketi, Grecia Salazar, Michael S. Ryoo, Krista Reymann, Kanishka Rao, Karl Pertsch, Igor Mordatch, Henryk Michalewski, Yao Lu, Sergey Levine, Lisa Lee, Tsang-Wei Edward Lee, Isabel Leal, Yuheng Kuang, Dmitry Kalashnikov, Ryan Julian, Nikhil J. Joshi, Alex Irpan, Brian Ichter, Jasmine Hsu, Alexander Herzog, Karol Hausman, Keerthana Gopalakrishnan, Chuyuan Fu, Pete Florence, Chelsea Finn, Kumar Avinava Dubey, Danny Driess, Tianli Ding, Krzysztof Marcin Choromanski, Xi Chen, Yevgen Chebotar, Justice Carbajal, Noah Brown, Anthony Brohan, Montserrat Gonzalez Arenas, Kehang Han:
RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control. CoRL 2023: 2165-2183 - [c371]Dhruv Shah, Michael Robert Equi, Blazej Osinski, Fei Xia, Brian Ichter, Sergey Levine:
Navigation with Large Language Models: Semantic Guesswork as a Heuristic for Planning. CoRL 2023: 2683-2699 - [c370]Kyle Stachowicz, Dhruv Shah, Arjun Bhorkar, Ilya Kostrikov, Sergey Levine:
FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing. CoRL 2023: 3100-3111 - [c369]Vivek Myers, Andre Wang He, Kuan Fang, Homer Rich Walke, Philippe Hansen-Estruch, Ching-An Cheng, Mihai Jalobeanu, Andrey Kolobov, Anca D. Dragan, Sergey Levine:
Goal Representations for Instruction Following: A Semi-Supervised Language Interface to Control. CoRL 2023: 3894-3908 - [c368]Yevgen Chebotar, Quan Vuong, Karol Hausman, Fei Xia, Yao Lu, Alex Irpan, Aviral Kumar, Tianhe Yu, Alexander Herzog, Karl Pertsch, Keerthana Gopalakrishnan, Julian Ibarz, Ofir Nachum, Sumedh Anand Sontakke, Grecia Salazar, Huong T. Tran, Jodilyn Peralta, Clayton Tan, Deeksha Manjunath, Jaspiar Singh, Brianna Zitkovich, Tomas Jackson, Kanishka Rao, Chelsea Finn, Sergey Levine:
Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions. CoRL 2023: 3909-3928 - [c367]Michael Chang, Alyssa L. Dayan, Franziska Meier, Thomas L. Griffiths, Sergey Levine, Amy Zhang:
Hierarchical Abstraction for Combinatorial Generalization in Object Rearrangement. ICLR 2023 - [c366]Raj Ghugare, Homanga Bharadhwaj, Benjamin Eysenbach, Sergey Levine, Russ Salakhutdinov:
Simplifying Model-based RL: Learning Representations, Latent-space Models, and Policies with One Objective. ICLR 2023 - [c365]Joey Hong, Aviral Kumar, Sergey Levine:
Confidence-Conditioned Value Functions for Offline Reinforcement Learning. ICLR 2023 - [c364]Aviral Kumar, Rishabh Agarwal, Xinyang Geng, George Tucker, Sergey Levine:
Offline Q-learning on Diverse Multi-Task Data Both Scales And Generalizes. ICLR 2023 - [c363]Qiyang Li, Aviral Kumar, Ilya Kostrikov, Sergey Levine:
Efficient Deep Reinforcement Learning Requires Regulating Overfitting. ICLR 2023 - [c362]Amrith Setlur, Don Kurian Dennis, Benjamin Eysenbach, Aditi Raghunathan, Chelsea Finn, Virginia Smith, Sergey Levine:
Bitrate-Constrained DRO: Beyond Worst Case Robustness To Unknown Group Shifts. ICLR 2023 - [c361]Charlie Snell, Ilya Kostrikov, Yi Su, Sherry Yang, Sergey Levine:
Offline RL for Natural Language Generation with Implicit Language Q Learning. ICLR 2023 - [c360]Philip J. Ball, Laura M. Smith, Ilya Kostrikov, Sergey Levine:
Efficient Online Reinforcement Learning with Offline Data. ICML 2023: 1577-1594 - [c359]Danny Driess, Fei Xia, Mehdi S. M. Sajjadi, Corey Lynch, Aakanksha Chowdhery, Brian Ichter, Ayzaan Wahid, Jonathan Tompson, Quan Vuong, Tianhe Yu, Wenlong Huang, Yevgen Chebotar, Pierre Sermanet, Daniel Duckworth, Sergey Levine, Vincent Vanhoucke, Karol Hausman, Marc Toussaint, Klaus Greff, Andy Zeng, Igor Mordatch, Pete Florence:
PaLM-E: An Embodied Multimodal Language Model. ICML 2023: 8469-8488 - [c358]Benjamin Eysenbach, Matthieu Geist, Sergey Levine, Ruslan Salakhutdinov:
A Connection between One-Step RL and Critic Regularization in Reinforcement Learning. ICML 2023: 9485-9507 - [c357]Dibya Ghosh, Chethan Anand Bhateja, Sergey Levine:
Reinforcement Learning from Passive Data via Latent Intentions. ICML 2023: 11321-11339 - [c356]Qiyang Li, Yuexiang Zhai, Yi Ma, Sergey Levine:
Understanding the Complexity Gains of Single-Task RL with a Curriculum. ICML 2023: 20412-20451 - [c355]Seohong Park, Sergey Levine:
Predictable MDP Abstraction for Unsupervised Model-Based RL. ICML 2023: 27246-27268 - [c354]Ikechukwu Uchendu, Ted Xiao, Yao Lu, Banghua Zhu, Mengyuan Yan, Joséphine Simon, Matthew Bennice, Chuyuan Fu, Cong Ma, Jiantao Jiao, Sergey Levine, Karol Hausman:
Jump-Start Reinforcement Learning. ICML 2023: 34556-34583 - [c353]Tony Tong Wang, Adam Gleave, Tom Tseng
, Kellin Pelrine, Nora Belrose, Joseph Miller, Michael D. Dennis, Yawen Duan, Viktor Pogrebniak, Sergey Levine, Stuart Russell:
Adversarial Policies Beat Superhuman Go AIs. ICML 2023: 35655-35739 - [c352]Noriaki Hirose, Dhruv Shah, Ajay Sridhar, Sergey Levine:
ExAug: Robot-Conditioned Navigation Policies via Geometric Experience Augmentation. ICRA 2023: 4077-4084 - [c351]Abhishek Gupta, Corey Lynch, Brandon Kinman
, Garrett Peake, Sergey Levine, Karol Hausman:
Demonstration-Bootstrapped Autonomous Practicing via Multi-Task Reinforcement Learning. ICRA 2023: 5020-5026 - [c350]Kelvin Xu, Zheyuan Hu
, Ria Doshi, Aaron Rovinsky, Vikash Kumar, Abhishek Gupta, Sergey Levine:
Dexterous Manipulation from Images: Autonomous Real-World RL via Substep Guidance. ICRA 2023: 5938-5945 - [c349]Ashvin Nair, Brian Zhu, Gokul Narayanan, Eugen Solowjow, Sergey Levine:
Learning on the Job: Self-Rewarding Offline-to-Online Finetuning for Industrial Insertion of Novel Connectors from Vision. ICRA 2023: 7154-7161 - [c348]Dhruv Shah, Ajay Sridhar, Arjun Bhorkar, Noriaki Hirose, Sergey Levine:
GNM: A General Navigation Model to Drive Any Robot. ICRA 2023: 7226-7233 - [c347]Jensen Gao, Siddharth Reddy, Glen Berseth, Anca D. Dragan, Sergey Levine:
Bootstrapping Adaptive Human-Machine Interfaces with Offline Reinforcement Learning. IROS 2023: 7523-7530 - [c346]Yiren Lu, Justin Fu, George Tucker, Xinlei Pan, Eli Bronstein, Rebecca Roelofs, Benjamin Sapp, Brandyn White, Aleksandra Faust
, Shimon Whiteson, Dragomir Anguelov, Sergey Levine:
Imitation Is Not Enough: Robustifying Imitation with Reinforcement Learning for Challenging Driving Scenarios. IROS 2023: 7553-7560 - [c345]Kyle Beltran Hatch, Benjamin Eysenbach, Rafael Rafailov, Tianhe Yu, Ruslan Salakhutdinov, Sergey Levine, Chelsea Finn:
Contrastive Example-Based Control. L4DC 2023: 155-169 - [c344]Thomas T. C. K. Zhang, Katie Kang, Bruce D. Lee, Claire J. Tomlin, Sergey Levine, Stephen Tu, Nikolai Matni:
Multi-Task Imitation Learning for Linear Dynamical Systems. L4DC 2023: 586-599 - [c343]Joey Hong, Sergey Levine, Anca D. Dragan:
Learning to Influence Human Behavior with Offline Reinforcement Learning. NeurIPS 2023 - [c342]Wenlong Huang, Fei Xia, Dhruv Shah, Danny Driess, Andy Zeng, Yao Lu, Pete Florence, Igor Mordatch, Sergey Levine, Karol Hausman, Brian Ichter:
Grounded Decoding: Guiding Text Generation with Grounded Models for Embodied Agents. NeurIPS 2023 - [c341]Qiyang Li, Jason Zhang, Dibya Ghosh, Amy Zhang, Sergey Levine:
Accelerating Exploration with Unlabeled Prior Data. NeurIPS 2023 - [c340]Mitsuhiko Nakamoto, Simon Zhai, Anikait Singh, Max Sobol Mark, Yi Ma, Chelsea Finn, Aviral Kumar, Sergey Levine:
Cal-QL: Calibrated Offline RL Pre-Training for Efficient Online Fine-Tuning. NeurIPS 2023 - [c339]Seohong Park, Dibya Ghosh, Benjamin Eysenbach, Sergey Levine:
HIQL: Offline Goal-Conditioned RL with Latent States as Actions. NeurIPS 2023 - [c338]Anikait Singh, Aviral Kumar, Quan Vuong, Yevgen Chebotar, Sergey Levine:
ReDS: Offline RL With Heteroskedastic Datasets via Support Constraints. NeurIPS 2023 - [c337]Manan Tomar, Riashat Islam, Matthew E. Taylor, Sergey Levine, Philip Bachman:
Ignorance is Bliss: Robust Control via Information Gating. NeurIPS 2023 - [c336]Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Joseph Dabis, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alexander Herzog, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Tomas Jackson, Sally Jesmonth, Nikhil J. Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Isabel Leal, Kuang-Huei Lee, Sergey Levine, Yao Lu, Utsav Malla, Deeksha Manjunath, Igor Mordatch, Ofir Nachum, Carolina Parada, Jodilyn Peralta, Emily Perez, Karl Pertsch, Jornell Quiambao, Kanishka Rao, Michael S. Ryoo, Grecia Salazar, Pannag R. Sanketi, Kevin Sayed, Jaspiar Singh, Sumedh Sontakke, Austin Stone, Clayton Tan, Huong T. Tran, Vincent Vanhoucke, Steve Vega, Quan Vuong, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Tianhe Yu, Brianna Zitkovich:
RT-1: Robotics Transformer for Real-World Control at Scale. Robotics: Science and Systems 2023 - [c335]Alexander Herzog, Kanishka Rao, Karol Hausman, Yao Lu, Paul Wohlhart, Mengyuan Yan, Jessica Lin, Montserrat Gonzalez Arenas, Ted Xiao, Daniel Kappler, Daniel Ho, Jarek Rettinghouse, Yevgen Chebotar, Kuang-Huei Lee, Keerthana Gopalakrishnan, Ryan Julian, Adrian Li, Chuyuan Fu, Bob Wei, Sangeetha Ramesh, Khem Holden, Kim Kleiven, David J. Rendleman, Sean Kirmani, Jeffrey Bingham, Jonathan Weisz, Ying Xu, Wenlong Lu, Matthew Bennice, Cody Fong, David Do, Jessica Lam, Yunfei Bai, Benjie Holson, Michael Quinlan, Noah Brown, Mrinal Kalakrishnan, Julian Ibarz, Peter Pastor, Sergey Levine:
Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators. Robotics: Science and Systems 2023 - [c334]Ilya Kostrikov, Laura M. Smith, Sergey Levine:
Demonstrating A Walk in the Park: Learning to Walk in 20 Minutes With Model-Free Reinforcement Learning. Robotics: Science and Systems 2023 - [c333]Aviral Kumar, Anikait Singh, Frederik D. Ebert, Mitsuhiko Nakamoto, Yanlai Yang, Chelsea Finn, Sergey Levine:
Pre-Training for Robots: Offline RL Enables Learning New Tasks in a Handful of Trials. Robotics: Science and Systems 2023 - [c332]Zhongyu Li, Xue Bin Peng, Pieter Abbeel, Sergey Levine, Glen Berseth
, Koushil Sreenath:
Robust and Versatile Bipedal Jumping Control through Reinforcement Learning. Robotics: Science and Systems 2023 - [c331]Laura M. Smith, J. Chase Kew, Tianyu Li, Linda Luu, Xue Bin Peng, Sehoon Ha, Jie Tan, Sergey Levine:
Learning and Adapting Agile Locomotion Skills by Transferring Experience. Robotics: Science and Systems 2023 - [c330]Ted Xiao, Harris Chan, Pierre Sermanet, Ayzaan Wahid, Anthony Brohan, Karol Hausman, Sergey Levine, Jonathan Tompson:
Robotic Skill Acquisition via Instruction Augmentation with Vision-Language Models. Robotics: Science and Systems 2023 - [c329]Tony Z. Zhao, Vikash Kumar, Sergey Levine, Chelsea Finn:
Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware. Robotics: Science and Systems 2023 - [e1]Alice Oh, Tristan Naumann, Amir Globerson, Kate Saenko, Moritz Hardt, Sergey Levine:
Advances in Neural Information Processing Systems 36: Annual Conference on Neural Information Processing Systems 2023, NeurIPS 2023, New Orleans, LA, USA, December 10 - 16, 2023. 2023 [contents] - [i427]Amrith Setlur, Don Kurian Dennis, Benjamin Eysenbach
, Aditi Raghunathan, Chelsea Finn, Virginia Smith, Sergey Levine:
Bitrate-Constrained DRO: Beyond Worst Case Robustness To Unknown Group Shifts. CoRR abs/2302.02931 (2023) - [i426]Philip J. Ball, Laura M. Smith, Ilya Kostrikov, Sergey Levine:
Efficient Online Reinforcement Learning with Offline Data. CoRR abs/2302.02948 (2023) - [i425]Seohong Park, Sergey Levine:
Predictable MDP Abstraction for Unsupervised Model-Based RL. CoRR abs/2302.03921 (2023) - [i424]Annie S. Chen, Yoonho Lee, Amrith Setlur, Sergey Levine, Chelsea Finn:
Project and Probe: Sample-Efficient Domain Adaptation by Interpolating Orthogonal Features. CoRR abs/2302.05441 (2023) - [i423]Zhongyu Li, Xue Bin Peng, Pieter Abbeel, Sergey Levine, Glen Berseth, Koushil Sreenath:
Robust and Versatile Bipedal Jumping Control through Multi-Task Reinforcement Learning. CoRR abs/2302.09450 (2023) - [i422]Wenlong Huang, Fei Xia, Dhruv Shah, Danny Driess, Andy Zeng, Yao Lu, Pete Florence, Igor Mordatch, Sergey Levine, Karol Hausman, Brian Ichter:
Grounded Decoding: Guiding Text Generation with Grounded Models for Robot Control. CoRR abs/2303.00855 (2023) - [i421]Joey Hong, Anca D. Dragan, Sergey Levine:
Learning to Influence Human Behavior with Offline Reinforcement Learning. CoRR abs/2303.02265 (2023) - [i420]Danny Driess, Fei Xia, Mehdi S. M. Sajjadi, Corey Lynch, Aakanksha Chowdhery, Brian Ichter, Ayzaan Wahid, Jonathan Tompson, Quan Vuong
, Tianhe Yu, Wenlong Huang, Yevgen Chebotar, Pierre Sermanet, Daniel Duckworth, Sergey Levine, Vincent Vanhoucke
, Karol Hausman, Marc Toussaint, Klaus Greff, Andy Zeng, Igor Mordatch, Pete Florence:
PaLM-E: An Embodied Multimodal Language Model. CoRR abs/2303.03378 (2023) - [i419]Mitsuhiko Nakamoto, Yuexiang Zhai, Anikait Singh, Max Sobol Mark, Yi Ma, Chelsea Finn, Aviral Kumar, Sergey Levine:
Cal-QL: Calibrated Offline RL Pre-Training for Efficient Online Fine-Tuning. CoRR abs/2303.05479 (2023) - [i418]Manan Tomar, Riashat Islam
, Sergey Levine, Philip Bachman:
Ignorance is Bliss: Robust Control via Information Gating. CoRR abs/2303.06121 (2023) - [i417]Michael Chang, Alyssa L. Dayan, Franziska Meier, Thomas L. Griffiths, Sergey Levine, Amy Zhang:
Neural Constraint Satisfaction: Hierarchical Abstraction for Combinatorial Generalization in Object Rearrangement. CoRR abs/2303.11373 (2023) - [i416]Dibya Ghosh, Chethan Bhateja, Sergey Levine:
Reinforcement Learning from Passive Data via Latent Intentions. CoRR abs/2304.04782 (2023) - [i415]Kyle Stachowicz, Dhruv Shah, Arjun Bhorkar, Ilya Kostrikov, Sergey Levine:
FastRLAP: A System for Learning High-Speed Driving via Deep RL and Autonomous Practicing. CoRR abs/2304.09831 (2023) - [i414]Laura M. Smith, J. Chase Kew, Tianyu Li, Linda Luu, Xue Bin Peng, Sehoon Ha, Jie Tan, Sergey Levine:
Learning and Adapting Agile Locomotion Skills by Transferring Experience. CoRR abs/2304.09834 (2023) - [i413]Qiyang Li, Aviral Kumar, Ilya Kostrikov, Sergey Levine:
Efficient Deep Reinforcement Learning Requires Regulating Overfitting. CoRR abs/2304.10466 (2023) - [i412]Philippe Hansen-Estruch, Ilya Kostrikov, Michael Janner, Jakub Grudzien Kuba, Sergey Levine:
IDQL: Implicit Q-Learning as an Actor-Critic Method with Diffusion Policies. CoRR abs/2304.10573 (2023) - [i411]Tony Z. Zhao, Vikash Kumar, Sergey Levine, Chelsea Finn:
Learning Fine-Grained Bimanual Manipulation with Low-Cost Hardware. CoRR abs/2304.13705 (2023) - [i410]Alexander Herzog, Kanishka Rao, Karol Hausman, Yao Lu, Paul Wohlhart, Mengyuan Yan, Jessica Lin, Montserrat Gonzalez Arenas, Ted Xiao, Daniel Kappler, Daniel Ho, Jarek Rettinghouse, Yevgen Chebotar, Kuang-Huei Lee, Keerthana Gopalakrishnan, Ryan Julian, Adrian Li, Chuyuan Kelly Fu, Bob Wei, Sangeetha Ramesh, Khem Holden, Kim Kleiven, David Rendleman, Sean Kirmani, Jeff Bingham, Jonathan Weisz, Ying Xu, Wenlong Lu, Matthew Bennice, Cody Fong, David Do, Jessica Lam, Yunfei Bai, Benjie Holson, Michael Quinlan, Noah Brown, Mrinal Kalakrishnan, Julian Ibarz, Peter Pastor, Sergey Levine:
Deep RL at Scale: Sorting Waste in Office Buildings with a Fleet of Mobile Manipulators. CoRR abs/2305.03270 (2023) - [i409]Kevin Black
, Michael Janner, Yilun Du, Ilya Kostrikov, Sergey Levine:
Training Diffusion Models with Reinforcement Learning. CoRR abs/2305.13301 (2023) - [i408]Arnav Gudibande, Eric Wallace, Charlie Snell, Xinyang Geng, Hao Liu, Pieter Abbeel, Sergey Levine, Dawn Song:
The False Promise of Imitating Proprietary LLMs. CoRR abs/2305.15717 (2023) - [i407]Noriaki Hirose, Dhruv Shah, Ajay Sridhar, Sergey Levine:
SACSoN: Scalable Autonomous Data Collection for Social Navigation. CoRR abs/2306.01874 (2023) - [i406]Chongyi Zheng, Benjamin Eysenbach, Homer Walke, Patrick Yin, Kuan Fang, Ruslan Salakhutdinov, Sergey Levine:
Stabilizing Contrastive RL: Techniques for Offline Goal Reaching. CoRR abs/2306.03346 (2023) - [i405]Annie S. Chen, Yoonho Lee, Amrith Setlur, Sergey Levine, Chelsea Finn:
Confidence-Based Model Selection: When to Take Shortcuts for Subpopulation Shifts. CoRR abs/2306.11120 (2023) - [i404]Dhruv Shah, Ajay Sridhar, Nitish Dashora, Kyle Stachowicz, Kevin Black
, Noriaki Hirose, Sergey Levine:
ViNT: A Foundation Model for Visual Navigation. CoRR abs/2306.14846 (2023) - [i403]Vivek Myers, Andre He, Kuan Fang, Homer Walke, Philippe Hansen-Estruch, Ching-An Cheng, Mihai Jalobeanu, Andrey Kolobov, Anca D. Dragan, Sergey Levine:
Goal Representations for Instruction Following: A Semi-Supervised Language Interface to Control. CoRR abs/2307.00117 (2023) - [i402]Jianlan Luo, Charles Xu, Xinyang Geng, Gilbert Feng, Kuan Fang, Liam Tan, Stefan Schaal, Sergey Levine:
Multi-Stage Cable Routing through Hierarchical Imitation Learning. CoRR abs/2307.08927 (2023) - [i401]Seohong Park, Dibya Ghosh, Benjamin Eysenbach
, Sergey Levine:
HIQL: Offline Goal-Conditioned RL with Latent States as Actions. CoRR abs/2307.11949 (2023) - [i400]Benjamin Eysenbach, Matthieu Geist, Sergey Levine, Ruslan Salakhutdinov:
A Connection between One-Step Regularization and Critic Regularization in Reinforcement Learning. CoRR abs/2307.12968 (2023) - [i399]Kyle Hatch, Benjamin Eysenbach, Rafael Rafailov, Tianhe Yu, Ruslan Salakhutdinov, Sergey Levine, Chelsea Finn:
Contrastive Example-Based Control. CoRR abs/2307.13101 (2023) - [i398]Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Xi Chen, Krzysztof Choromanski, Tianli Ding, Danny Driess, Avinava Dubey, Chelsea Finn, Pete Florence, Chuyuan Fu, Montse Gonzalez Arenas, Keerthana Gopalakrishnan, Kehang Han, Karol Hausman, Alexander Herzog, Jasmine Hsu, Brian Ichter, Alex Irpan, Nikhil J. Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Isabel Leal, Lisa Lee, Tsang-Wei Edward Lee, Sergey Levine, Yao Lu, Henryk Michalewski, Igor Mordatch, Karl Pertsch, Kanishka Rao, Krista Reymann, Michael S. Ryoo, Grecia Salazar, Pannag Sanketi, Pierre Sermanet, Jaspiar Singh, Anikait Singh, Radu Soricut, Huong T. Tran, Vincent Vanhoucke
, Quan Vuong
, Ayzaan Wahid, Stefan Welker, Paul Wohlhart, Jialin Wu, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Tianhe Yu, Brianna Zitkovich:
RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control. CoRR abs/2307.15818 (2023) - [i397]Homer Walke, Kevin Black
, Abraham Lee, Moo Jin Kim, Maximilian Du, Chongyi Zheng, Tony Z. Zhao, Philippe Hansen-Estruch, Quan Vuong
, Andre He, Vivek Myers, Kuan Fang, Chelsea Finn, Sergey Levine:
BridgeData V2: A Dataset for Robot Learning at Scale. CoRR abs/2308.12952 (2023) - [i396]Zheyuan Hu, Aaron Rovinsky, Jianlan Luo, Vikash Kumar, Abhishek Gupta, Sergey Levine:
REBOOT: Reuse Data for Bootstrapping Efficient Real-World Dexterous Manipulation. CoRR abs/2309.03322 (2023) - [i395]Jensen Gao, Siddharth Reddy, Glen Berseth, Anca D. Dragan, Sergey Levine:
Bootstrapping Adaptive Human-Machine Interfaces with Offline Reinforcement Learning. CoRR abs/2309.03839 (2023) - [i394]Yevgen Chebotar, Quan Vuong
, Alex Irpan, Karol Hausman, Fei Xia, Yao Lu, Aviral Kumar, Tianhe Yu, Alexander Herzog, Karl Pertsch, Keerthana Gopalakrishnan, Julian Ibarz, Ofir Nachum, Sumedh Sontakke, Grecia Salazar, Huong T. Tran, Jodilyn Peralta, Clayton Tan, Deeksha Manjunath, Jaspiar Singh, Brianna Zitkovich, Tomas Jackson, Kanishka Rao, Chelsea Finn, Sergey Levine:
Q-Transformer: Scalable Offline Reinforcement Learning via Autoregressive Q-Functions. CoRR abs/2309.10150 (2023) - [i393]Chethan Bhateja, Derek Guo, Dibya Ghosh, Anikait Singh, Manan Tomar, Quan Vuong
, Yevgen Chebotar, Sergey Levine, Aviral Kumar:
Robotic Offline RL from Internet Videos via Value-Function Pre-Training. CoRR abs/2309.13041 (2023) - [i392]Katie Kang, Amrith Setlur, Claire J. Tomlin, Sergey Levine:
Deep Neural Networks Tend To Extrapolate Predictably. CoRR abs/2310.00873 (2023) - [i391]Ajay Sridhar, Dhruv Shah, Catherine Glossop, Sergey Levine:
NoMaD: Goal Masked Diffusion Policies for Navigation and Exploration. CoRR abs/2310.07896 (2023) - [i390]Max Sobol Mark, Archit Sharma, Fahim Tajwar, Rafael Rafailov, Sergey Levine, Chelsea Finn:
Offline Retraining for Online RL: Decoupled Policy Learning to Mitigate Exploration Bias. CoRR abs/2310.08558 (2023) - [i389]Seohong Park, Oleh Rybkin, Sergey Levine:
METRA: Scalable Unsupervised RL with Metric-Aware Abstraction. CoRR abs/2310.08887 (2023) - [i388]Han Qi, Xinyang Geng, Stefano Rando, Iku Ohama, Aviral Kumar, Sergey Levine:
Latent Conservative Objective Models for Data-Driven Crystal Structure Prediction. CoRR abs/2310.10056 (2023) - [i387]Dhruv Shah, Michael Equi, Blazej Osinski, Fei Xia, Brian Ichter, Sergey Levine:
Navigation with Large Language Models: Semantic Guesswork as a Heuristic for Planning. CoRR abs/2310.10103 (2023) - [i386]Kevin Black
, Mitsuhiko Nakamoto, Pranav Atreya, Homer Walke, Chelsea Finn, Aviral Kumar, Sergey Levine:
Zero-Shot Robotic Manipulation with Pretrained Image-Editing Diffusion Models. CoRR abs/2310.10639 (2023) - [i385]Jianlan Luo, Perry Dong, Jeffrey Wu, Aviral Kumar, Xinyang Geng, Sergey Levine:
Action-Quantized Offline Reinforcement Learning for Robotic Skill Learning. CoRR abs/2310.11731 (2023) - [i384]Laura M. Smith, Yunhao Cao, Sergey Levine:
Grow Your Limits: Continuous Improvement with Real-World RL for Robotic Locomotion. CoRR abs/2310.17634 (2023) - [i383]Joey Hong, Anca D. Dragan, Sergey Levine:
Offline RL with Observation Histories: Analyzing and Improving Sample Complexity. CoRR abs/2310.20663 (2023) - [i382]Annie S. Chen, Govind Chada, Laura M. Smith, Archit Sharma, Zipeng Fu, Sergey Levine, Chelsea Finn:
Adapt On-the-Go: Behavior Modulation for Single-Life Robot Deployment. CoRR abs/2311.01059 (2023) - [i381]Qiyang Li, Jason Zhang, Dibya Ghosh, Amy Zhang, Sergey Levine:
Accelerating Exploration with Unlabeled Prior Data. CoRR abs/2311.05067 (2023) - [i380]Joey Hong, Sergey Levine, Anca D. Dragan:
Zero-Shot Goal-Directed Dialogue via RL on Imagined Conversations. CoRR abs/2311.05584 (2023) - [i379]Jianlan Luo, Perry Dong, Yuexiang Zhai, Yi Ma, Sergey Levine:
RLIF: Interactive Imitation Learning as Reinforcement Learning. CoRR abs/2311.12996 (2023) - [i378]Marwa Abdulhai, Isadora White, Charlie Snell, Charles Sun, Joey Hong, Yuexiang Zhai, Kelvin Xu, Sergey Levine:
LMRL Gym: Benchmarks for Multi-Turn Reinforcement Learning with Language Models. CoRR abs/2311.18232 (2023) - [i377]Chengshu Li, Jacky Liang, Andy Zeng, Xinyun Chen, Karol Hausman, Dorsa Sadigh, Sergey Levine, Li Fei-Fei, Fei Xia, Brian Ichter:
Chain of Code: Reasoning with a Language Model-Augmented Code Emulator. CoRR abs/2312.04474 (2023) - 2022
- [j20]Xue Bin Peng, Yunrong Guo, Lina Halper, Sergey Levine, Sanja Fidler:
ASE: large-scale reusable adversarial skill embeddings for physically simulated characters. ACM Trans. Graph. 41(4): 94:1-94:17 (2022) - [c328]Dhruv Shah, Arjun Bhorkar, Hrishit Leen, Ilya Kostrikov, Nicholas Rhinehart, Sergey Levine:
Offline Reinforcement Learning for Visual Navigation. CoRL 2022: 44-54 - [c327]Kuan Fang, Patrick Yin, Ashvin Nair, Homer Walke, Gengchen Yan, Sergey Levine:
Generalization with Lossy Affordances: Leveraging Broad Offline Data for Learning Visuomotor Tasks. CoRL 2022: 106-117 - [c326]Brian Ichter, Anthony Brohan, Yevgen Chebotar, Chelsea Finn, Karol Hausman, Alexander Herzog, Daniel Ho, Julian Ibarz, Alex Irpan, Eric Jang, Ryan Julian, Dmitry Kalashnikov, Sergey Levine, Yao Lu, Carolina Parada, Kanishka Rao, Pierre Sermanet, Alexander Toshev, Vincent Vanhoucke, Fei Xia, Ted Xiao, Peng Xu, Mengyuan Yan, Noah Brown, Michael Ahn, Omar Cortes, Nicolas Sievers, Clayton Tan, Sichun Xu, Diego Reyes, Jarek Rettinghouse, Jornell Quiambao, Peter Pastor, Linda Luu, Kuang-Huei Lee, Yuheng Kuang, Sally Jesmonth, Nikhil J. Joshi, Kyle Jeffrey, Rosario Jauregui Ruano, Jasmine Hsu, Keerthana Gopalakrishnan, Byron David, Andy Zeng, Chuyuan Kelly Fu:
Do As I Can, Not As I Say: Grounding Language in Robotic Affordances. CoRL 2022: 287-318 - [c325]Dhruv Shah, Blazej Osinski, Brian Ichter, Sergey Levine:
LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action. CoRL 2022: 492-504 - [c324]Charles Packer, Nicholas Rhinehart, Rowan Thomas McAllister, Matthew A. Wright, Xin Wang, Jeff He, Sergey Levine, Joseph E. Gonzalez:
Is Anyone There? Learning a Planner Contingent on Perceptual Uncertainty. CoRL 2022: 1607-1617 - [c323]Homer Walke, Jonathan Yang, Albert Yu, Aviral Kumar, Jedrzej Orbik, Avi Singh, Sergey Levine:
Don't Start From Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning. CoRL 2022: 1652-1662 - [c322]Wenlong Huang, Fei Xia, Ted Xiao, Harris Chan, Jacky Liang, Pete Florence, Andy Zeng, Jonathan Tompson, Igor Mordatch, Yevgen Chebotar, Pierre Sermanet, Tomas Jackson, Noah Brown, Linda Luu, Sergey Levine, Karol Hausman, Brian Ichter:
Inner Monologue: Embodied Reasoning through Planning with Language Models. CoRL 2022: 1769-1782 - [c321]Gilbert Feng, Hongbo Zhang, Zhongyu Li, Xue Bin Peng, Bhuvan Basireddy, Linzhu Yue, Zhitao Song, Lizhi Yang, Yunhui Liu, Koushil Sreenath, Sergey Levine:
GenLoco: Generalized Locomotion Controllers for Quadrupedal Robots. CoRL 2022: 1893-1903 - [c320]Glen Berseth
, Zhiwei Zhang, Grace Zhang, Chelsea Finn, Sergey Levine:
CoMPS: Continual Meta Policy Search. ICLR 2022 - [c319]Homanga Bharadhwaj, Mohammad Babaeizadeh, Dumitru Erhan, Sergey Levine:
Information Prioritization through Empowerment in Visual Model-based RL. ICLR 2022 - [c318]Scott Emmons, Benjamin Eysenbach, Ilya Kostrikov, Sergey Levine:
RvS: What is Essential for Offline RL via Supervised Learning? ICLR 2022 - [c317]Benjamin Eysenbach, Sergey Levine:
Maximum Entropy RL (Provably) Solves Some Robust RL Problems. ICLR 2022 - [c316]Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine:
The Information Geometry of Unsupervised Reinforcement Learning. ICLR 2022 - [c315]Ilya Kostrikov, Ashvin Nair, Sergey Levine:
Offline Reinforcement Learning with Implicit Q-Learning. ICLR 2022 - [c314]Aviral Kumar, Rishabh Agarwal, Tengyu Ma, Aaron C. Courville, George Tucker, Sergey Levine:
DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization. ICLR 2022 - [c313]Aviral Kumar, Joey Hong, Anikait Singh, Sergey Levine:
Should I Run Offline Reinforcement Learning or Behavioral Cloning? ICLR 2022 - [c312]Aviral Kumar, Amir Yazdanbakhsh, Milad Hashemi, Kevin Swersky, Sergey Levine:
Data-Driven Offline Optimization for Architecting Hardware Accelerators. ICLR 2022 - [c311]Shiori Sagawa, Pang Wei Koh, Tony Lee, Irena Gao, Sang Michael Xie, Kendrick Shen, Ananya Kumar, Weihua Hu, Michihiro Yasunaga, Henrik Marklund, Sara Beery, Etienne David, Ian Stavness, Wei Guo
, Jure Leskovec, Kate Saenko, Tatsunori Hashimoto, Sergey Levine, Chelsea Finn, Percy Liang:
Extending the WILDS Benchmark for Unsupervised Adaptation. ICLR 2022 - [c310]Dhruv Shah, Peng Xu, Yao Lu, Ted Xiao, Alexander Toshev, Sergey Levine, Brian Ichter:
Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning. ICLR 2022 - [c309]Archit Sharma, Kelvin Xu, Nikhil Sardana, Abhishek Gupta, Karol Hausman, Sergey Levine, Chelsea Finn:
Autonomous Reinforcement Learning: Formalism and Benchmarking. ICLR 2022 - [c308]Mengjiao Yang, Sergey Levine, Ofir Nachum:
TRAIL: Near-Optimal Imitation Learning with Suboptimal Data. ICLR 2022 - [c307]Tianjun Zhang, Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine, Joseph E. Gonzalez:
C-Planning: An Automatic Curriculum for Learning Goal-Reaching Tasks. ICLR 2022 - [c306]Dibya Ghosh, Anurag Ajay, Pulkit Agrawal, Sergey Levine:
Offline RL Policies Should Be Trained to be Adaptive. ICML 2022: 7513-7530 - [c305]Philippe Hansen-Estruch, Amy Zhang, Ashvin Nair, Patrick Yin, Sergey Levine:
Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning. ICML 2022: 8407-8426 - [c304]Michael Janner, Yilun Du, Joshua B. Tenenbaum, Sergey Levine:
Planning with Diffusion for Flexible Behavior Synthesis. ICML 2022: 9902-9915 - [c303]Katie Kang, Paula Gradu, Jason J. Choi, Michael Janner, Claire J. Tomlin, Sergey Levine:
Lyapunov Density Models: Constraining Distribution Shift in Learning-Based Control. ICML 2022: 10708-10733 - [c302]Vitchyr H. Pong, Ashvin Nair, Laura M. Smith, Catherine Huang, Sergey Levine:
Offline Meta-Reinforcement Learning with Online Self-Supervision. ICML 2022: 17811-17829 - [c301]Brandon Trabucco, Xinyang Geng, Aviral Kumar, Sergey Levine:
Design-Bench: Benchmarks for Data-Driven Offline Model-Based Optimization. ICML 2022: 21658-21676 - [c300]Tianhe Yu, Aviral Kumar, Yevgen Chebotar, Karol Hausman, Chelsea Finn, Sergey Levine:
How to Leverage Unlabeled Data in Offline Reinforcement Learning. ICML 2022: 25611-25635 - [c299]Rowan McAllister, Blake Wulfe, Jean Mercat, Logan Ellis, Sergey Levine, Adrien Gaidon:
Control-Aware Prediction Objectives for Autonomous Driving. ICRA 2022: 1-8 - [c298]Laura M. Smith, J. Chase Kew, Xue Bin Peng, Sehoon Ha, Jie Tan, Sergey Levine:
Legged Robots that Keep on Learning: Fine-Tuning Locomotion Policies in the Real World. ICRA 2022: 1593-1599 - [c297]Nitish Dashora, Daniel Shin, Dhruv Shah, Henry A. Leopold, David D. Fan, Ali-Akbar Agha-Mohammadi, Nicholas Rhinehart, Sergey Levine:
Hybrid Imitative Planning with Geometric and Predictive Costs in Off-road Environments. ICRA 2022: 4452-4458 - [c296]Tony Z. Zhao, Jianlan Luo, Oleg Sushkov, Rugile Pevceviciute, Nicolas Heess, Jon Scholz, Stefan Schaal, Sergey Levine:
Offline Meta-Reinforcement Learning for Industrial Insertion. ICRA 2022: 6386-6393 - [c295]Sean Chen, Jensen Gao, Siddharth Reddy, Glen Berseth, Anca D. Dragan, Sergey Levine:
ASHA: Assistive Teleoperation via Human-in-the-Loop Reinforcement Learning. ICRA 2022: 7505-7512 - [c294]Yandong Ji
, Zhongyu Li, Yinan Sun, Xue Bin Peng, Sergey Levine, Glen Berseth, Koushil Sreenath
:
Hierarchical Reinforcement Learning for Precise Soccer Shooting Skills using a Quadrupedal Robot. IROS 2022: 1479-1486 - [c293]Kuan Fang, Patrick Yin, Ashvin Nair, Sergey Levine:
Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space. IROS 2022: 4076-4083 - [c292]Charlie Snell, Sherry Yang, Justin Fu, Yi Su, Sergey Levine:
Context-Aware Language Modeling for Goal-Oriented Dialogue Systems. NAACL-HLT (Findings) 2022: 2351-2366 - [c291]Siddharth Verma, Justin Fu, Sherry Yang, Sergey Levine:
CHAI: A CHatbot AI for Task-Oriented Dialogue with Offline Reinforcement Learning. NAACL-HLT 2022: 4471-4491 - [c290]Michael Chang, Tom Griffiths, Sergey Levine:
Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation. NeurIPS 2022 - [c289]Abhishek Gupta, Aldo Pacchiano, Yuexiang Zhai, Sham M. Kakade, Sergey Levine:
Unpacking Reward Shaping: Understanding the Benefits of Reward Engineering on Sample Complexity. NeurIPS 2022 - [c288]Anurag Ajay, Abhishek Gupta, Dibya Ghosh, Sergey Levine, Pulkit Agrawal:
Distributionally Adaptive Meta Reinforcement Learning. NeurIPS 2022 - [c287]Annie S. Chen, Archit Sharma, Sergey Levine, Chelsea Finn:
You Only Live Once: Single-Life Reinforcement Learning. NeurIPS 2022 - [c286]Benjamin Eysenbach, Alexander Khazatsky, Sergey Levine, Ruslan Salakhutdinov:
Mismatched No More: Joint Model-Policy Optimization for Model-Based RL. NeurIPS 2022 - [c285]Benjamin Eysenbach, Soumith Udatha, Russ Salakhutdinov, Sergey Levine:
Imitating Past Successes can be Very Suboptimal. NeurIPS 2022 - [c284]Benjamin Eysenbach, Tianjun Zhang, Sergey Levine, Ruslan Salakhutdinov:
Contrastive Learning as Goal-Conditioned Reinforcement Learning. NeurIPS 2022 - [c283]Han Qi, Yi Su, Aviral Kumar, Sergey Levine:
Data-Driven Offline Decision-Making via Invariant Representation Learning. NeurIPS 2022 - [c282]Siddharth Reddy, Sergey Levine, Anca D. Dragan:
First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization. NeurIPS 2022 - [c281]Amrith Setlur, Benjamin Eysenbach, Virginia Smith, Sergey Levine:
Adversarial Unlearning: Reducing Confidence Along Adversarial Directions. NeurIPS 2022 - [c280]Quan Vuong, Aviral Kumar, Sergey Levine, Yevgen Chebotar:
DASCO: Dual-Generator Adversarial Support Constrained Offline Reinforcement Learning. NeurIPS 2022 - [c279]Marvin Zhang, Sergey Levine, Chelsea Finn:
MEMO: Test Time Robustness via Adaptation and Augmentation. NeurIPS 2022 - [c278]Frederik Ebert, Yanlai Yang, Karl Schmeckpeper, Bernadette Bucher
, Georgios Georgakis, Kostas Daniilidis, Chelsea Finn, Sergey Levine:
Bridge Data: Boosting Generalization of Robotic Skills with Cross-Domain Datasets. Robotics: Science and Systems 2022 - [c277]Dhruv Shah, Sergey Levine:
ViKiNG: Vision-Based Kilometer-Scale Navigation with Geographic Hints. Robotics: Science and Systems 2022 - [i376]Jathushan Rajasegaran, Chelsea Finn, Sergey Levine:
Fully Online Meta-Learning Without Task Boundaries. CoRR abs/2202.00263 (2022) - [i375]Tianhe Yu, Aviral Kumar, Yevgen Chebotar, Karol Hausman, Chelsea Finn, Sergey Levine:
How to Leverage Unlabeled Data in Offline Reinforcement Learning. CoRR abs/2202.01741 (2022) - [i374]Eric Jang, Alex Irpan, Mohi Khansari, Daniel Kappler, Frederik Ebert, Corey Lynch, Sergey Levine, Chelsea Finn:
BC-Z: Zero-Shot Task Generalization with Robotic Imitation Learning. CoRR abs/2202.02005 (2022) - [i373]Sean Chen, Jensen Gao, Siddharth Reddy, Glen Berseth, Anca D. Dragan, Sergey Levine:
ASHA: Assistive Teleoperation via Human-in-the-Loop Reinforcement Learning. CoRR abs/2202.02465 (2022) - [i372]Brandon Trabucco, Xinyang Geng, Aviral Kumar, Sergey Levine:
Design-Bench: Benchmarks for Data-Driven Offline Model-Based Optimization. CoRR abs/2202.08450 (2022) - [i371]Dhruv Shah, Sergey Levine:
ViKiNG: Vision-Based Kilometer-Scale Navigation with Geographic Hints. CoRR abs/2202.11271 (2022) - [i370]Jensen Gao, Siddharth Reddy, Glen Berseth, Nicholas Hardy, Nikhilesh Natraj, Karunesh Ganguly, Anca D. Dragan, Sergey Levine:
X2T: Training an X-to-Text Typing Interface with Online Learning from User Feedback. CoRR abs/2203.02072 (2022) - [i369]Abhishek Gupta, Corey Lynch, Brandon Kinman, Garrett Peake, Sergey Levine, Karol Hausman:
Demonstration-Bootstrapped Autonomous Practicing via Multi-Task Reinforcement Learning. CoRR abs/2203.15755 (2022) - [i368]Michael Ahn, Anthony Brohan, Noah Brown, Yevgen Chebotar, Omar Cortes, Byron David, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alexander Herzog, Daniel Ho, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Eric Jang, Rosario Jauregui Ruano, Kyle Jeffrey, Sally Jesmonth, Nikhil J. Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Kuang-Huei Lee, Sergey Levine, Yao Lu, Linda Luu, Carolina Parada, Peter Pastor, Jornell Quiambao, Kanishka Rao, Jarek Rettinghouse, Diego Reyes, Pierre Sermanet, Nicolas Sievers, Clayton Tan, Alexander Toshev, Vincent Vanhoucke
, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Mengyuan Yan:
Do As I Can, Not As I Say: Grounding Language in Robotic Affordances. CoRR abs/2204.01691 (2022) - [i367]Ikechukwu Uchendu, Ted Xiao, Yao Lu, Banghua Zhu, Mengyuan Yan, Joséphine Simon, Matthew Bennice, Chuyuan Fu, Cong Ma, Jiantao Jiao, Sergey Levine, Karol Hausman:
Jump-Start Reinforcement Learning. CoRR abs/2204.02372 (2022) - [i366]Aviral Kumar, Joey Hong, Anikait Singh, Sergey Levine:
When Should We Prefer Offline Reinforcement Learning Over Behavioral Cloning? CoRR abs/2204.05618 (2022) - [i365]Siddharth Verma, Justin Fu, Mengjiao Yang, Sergey Levine:
CHAI: A CHatbot AI for Task-Oriented Dialogue with Offline Reinforcement Learning. CoRR abs/2204.08426 (2022) - [i364]Homanga Bharadhwaj, Mohammad Babaeizadeh, Dumitru Erhan, Sergey Levine:
INFOrmation Prioritization through EmPOWERment in Visual Model-Based RL. CoRR abs/2204.08585 (2022) - [i363]Charlie Snell, Mengjiao Yang, Justin Fu, Yi Su, Sergey Levine:
Context-Aware Language Modeling for Goal-Oriented Dialogue Systems. CoRR abs/2204.10198 (2022) - [i362]Philippe Hansen-Estruch, Amy Zhang, Ashvin Nair, Patrick Yin, Sergey Levine:
Bisimulation Makes Analogies in Goal-Conditioned Reinforcement Learning. CoRR abs/2204.13060 (2022) - [i361]Rowan McAllister, Blake Wulfe, Jean Mercat, Logan Ellis, Sergey Levine, Adrien Gaidon:
Control-Aware Prediction Objectives for Autonomous Driving. CoRR abs/2204.13319 (2022) - [i360]Xue Bin Peng, Yunrong Guo, Lina Halper, Sergey Levine, Sanja Fidler:
ASE: Large-Scale Reusable Adversarial Skill Embeddings for Physically Simulated Characters. CoRR abs/2205.01906 (2022) - [i359]Kuan Fang, Patrick Yin, Ashvin Nair, Sergey Levine:
Planning to Practice: Efficient Online Fine-Tuning by Composing Goals in Latent Space. CoRR abs/2205.08129 (2022) - [i358]Michael Janner, Yilun Du, Joshua B. Tenenbaum, Sergey Levine:
Planning with Diffusion for Flexible Behavior Synthesis. CoRR abs/2205.09991 (2022) - [i357]Siddharth Reddy, Sergey Levine, Anca D. Dragan:
First Contact: Unsupervised Human-Machine Co-Adaptation via Mutual Information Maximization. CoRR abs/2205.12381 (2022) - [i356]Xinyang Geng, Hao Liu, Lisa Lee, Dale Schuurams, Sergey Levine, Pieter Abbeel:
Multimodal Masked Autoencoders Learn Transferable Representations. CoRR abs/2205.14204 (2022) - [i355]Amrith Setlur, Benjamin Eysenbach
, Virginia Smith, Sergey Levine:
Adversarial Unlearning: Reducing Confidence Along Adversarial Directions. CoRR abs/2206.01367 (2022) - [i354]Benjamin Eysenbach
, Soumith Udatha, Sergey Levine, Ruslan Salakhutdinov:
Imitating Past Successes can be Very Suboptimal. CoRR abs/2206.03378 (2022) - [i353]Benjamin Eysenbach
, Tianjun Zhang, Ruslan Salakhutdinov, Sergey Levine:
Contrastive Learning as Goal-Conditioned Reinforcement Learning. CoRR abs/2206.07568 (2022) - [i352]Katie Kang, Paula Gradu, Jason J. Choi, Michael Janner, Claire J. Tomlin, Sergey Levine:
Lyapunov Density Models: Constraining Distribution Shift in Learning-Based Control. CoRR abs/2206.10524 (2022) - [i351]Charlie Snell, Ilya Kostrikov, Yi Su, Mengjiao Yang, Sergey Levine:
Offline RL for Natural Language Generation with Implicit Language Q Learning. CoRR abs/2206.11871 (2022) - [i350]Michael Chang, Thomas L. Griffiths, Sergey Levine:
Object Representations as Fixed Points: Training Iterative Refinement Algorithms with Implicit Differentiation. CoRR abs/2207.00787 (2022) - [i349]Dibya Ghosh, Anurag Ajay, Pulkit Agrawal, Sergey Levine:
Offline RL Policies Should be Trained to be Adaptive. CoRR abs/2207.02200 (2022) - [i348]Dhruv Shah, Blazej Osinski, Brian Ichter, Sergey Levine:
LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action. CoRR abs/2207.04429 (2022) - [i347]Homer Walke, Jonathan Yang, Albert Yu, Aviral Kumar, Jedrzej Orbik, Avi Singh, Sergey Levine:
Don't Start From Scratch: Leveraging Prior Data to Automate Robotic Reinforcement Learning. CoRR abs/2207.04703 (2022) - [i346]Wenlong Huang, Fei Xia, Ted Xiao, Harris Chan, Jacky Liang, Pete Florence, Andy Zeng, Jonathan Tompson, Igor Mordatch, Yevgen Chebotar, Pierre Sermanet, Noah Brown, Tomas Jackson, Linda Luu, Sergey Levine, Karol Hausman, Brian Ichter:
Inner Monologue: Embodied Reasoning through Planning with Language Models. CoRR abs/2207.05608 (2022) - [i345]Yandong Ji
, Zhongyu Li, Yinan Sun, Xue Bin Peng, Sergey Levine, Glen Berseth, Koushil Sreenath:
Hierarchical Reinforcement Learning for Precise Soccer Shooting Skills using a Quadrupedal Robot. CoRR abs/2208.01160 (2022) - [i344]Marwa Abdulhai, Natasha Jaques, Sergey Levine:
Basis for Intentions: Efficient Inverse Reinforcement Learning using Past Experience. CoRR abs/2208.04919 (2022) - [i343]Laura M. Smith, Ilya Kostrikov, Sergey Levine:
A Walk in the Park: Learning to Walk in 20 Minutes With Model-Free Reinforcement Learning. CoRR abs/2208.07860 (2022) - [i342]Gilbert Feng, Hongbo Zhang, Zhongyu Li, Xue Bin Peng, Bhuvan Basireddy, Linzhu Yue, Zhitao Song, Lizhi Yang, Yunhui Liu, Koushil Sreenath, Sergey Levine:
GenLoco: Generalized Locomotion Controllers for Quadrupedal Robots. CoRR abs/2209.05309 (2022) - [i341]Raj Ghugare, Homanga Bharadhwaj, Benjamin Eysenbach
, Sergey Levine, Ruslan Salakhutdinov:
Simplifying Model-based RL: Learning Representations, Latent-space Models, and Policies with One Objective. CoRR abs/2209.08466 (2022) - [i340]Anurag Ajay, Abhishek Gupta, Dibya Ghosh, Sergey Levine, Pulkit Agrawal:
Distributionally Adaptive Meta Reinforcement Learning. CoRR abs/2210.03104 (2022) - [i339]Dhruv Shah, Ajay Sridhar, Arjun Bhorkar, Noriaki Hirose, Sergey Levine:
GNM: A General Navigation Model to Drive Any Robot. CoRR abs/2210.03370 (2022) - [i338]Aviral Kumar, Anikait Singh, Frederik Ebert, Yanlai Yang, Chelsea Finn, Sergey Levine:
Pre-Training for Robots: Offline RL Enables Learning New Tasks from a Handful of Trials. CoRR abs/2210.05178 (2022) - [i337]Kuan Fang, Patrick Yin, Ashvin Nair, Homer Walke, Gengchen Yan, Sergey Levine:
Generalization with Lossy Affordances: Leveraging Broad Offline Data for Learning Visuomotor Tasks. CoRR abs/2210.06601 (2022) - [i336]Noriaki Hirose, Dhruv Shah, Ajay Sridhar, Sergey Levine:
ExAug: Robot-Conditioned Navigation Policies via Geometric Experience Augmentation. CoRR abs/2210.07450 (2022) - [i335]Annie S. Chen, Archit Sharma, Sergey Levine, Chelsea Finn:
You Only Live Once: Single-Life Reinforcement Learning. CoRR abs/2210.08863 (2022) - [i334]Abhishek Gupta, Aldo Pacchiano, Yuexiang Zhai, Sham M. Kakade, Sergey Levine:
Unpacking Reward Shaping: Understanding the Benefits of Reward Engineering on Sample Complexity. CoRR abs/2210.09579 (2022) - [i333]Hao Liu, Xinyang Geng, Lisa Lee, Igor Mordatch, Sergey Levine, Sharan Narang, Pieter Abbeel:
FCM: Forgetful Causal Masking Makes Causal Language Models Better Zero-Shot Learners. CoRR abs/2210.13432 (2022) - [i332]Ashvin Nair, Brian Zhu, Gokul Narayanan, Eugen Solowjow, Sergey Levine:
Learning on the Job: Self-Rewarding Offline-to-Online Finetuning for Industrial Insertion of Novel Connectors from Vision. CoRR abs/2210.15206 (2022) - [i331]Tony Tong Wang, Adam Gleave
, Nora Belrose, Tom Tseng, Joseph Miller, Michael D. Dennis, Yawen Duan, Viktor Pogrebniak, Sergey Levine, Stuart Russell:
Adversarial Policies Beat Professional-Level Go AIs. CoRR abs/2211.00241 (2022) - [i330]Anikait Singh, Aviral Kumar, Quan Vuong
, Yevgen Chebotar, Sergey Levine:
Offline RL With Realistic Datasets: Heteroskedasticity and Support Constraints. CoRR abs/2211.01052 (2022) - [i329]Quan Vuong
, Aviral Kumar, Sergey Levine, Yevgen Chebotar:
Dual Generator Offline Reinforcement Learning. CoRR abs/2211.01471 (2022) - [i328]Han Qi, Yi Su, Aviral Kumar, Sergey Levine:
Data-Driven Offline Decision-Making via Invariant Representation Learning. CoRR abs/2211.11349 (2022) - [i327]Ted Xiao, Harris Chan, Pierre Sermanet, Ayzaan Wahid, Anthony Brohan, Karol Hausman, Sergey Levine, Jonathan Tompson:
Robotic Skill Acquisition via Instruction Augmentation with Vision-Language Models. CoRR abs/2211.11736 (2022) - [i326]Aviral Kumar, Rishabh Agarwal, Xinyang Geng, George Tucker, Sergey Levine:
Offline Q-Learning on Diverse Multi-Task Data Both Scales And Generalizes. CoRR abs/2211.15144 (2022) - [i325]Thomas T. C. K. Zhang, Katie Kang, Bruce D. Lee, Claire J. Tomlin, Sergey Levine, Stephen Tu, Nikolai Matni:
Multi-Task Imitation Learning for Linear Dynamical Systems. CoRR abs/2212.00186 (2022) - [i324]Joey Hong, Aviral Kumar, Sergey Levine:
Confidence-Conditioned Value Functions for Offline Reinforcement Learning. CoRR abs/2212.04607 (2022) - [i323]Sergey Levine, Dhruv Shah:
Learning Robotic Navigation from Experience: Principles, Methods, and Recent Results. CoRR abs/2212.06759 (2022) - [i322]Anthony Brohan, Noah Brown, Justice Carbajal, Yevgen Chebotar, Joseph Dabis, Chelsea Finn, Keerthana Gopalakrishnan, Karol Hausman, Alexander Herzog, Jasmine Hsu, Julian Ibarz, Brian Ichter, Alex Irpan, Tomas Jackson, Sally Jesmonth, Nikhil J. Joshi, Ryan Julian, Dmitry Kalashnikov, Yuheng Kuang, Isabel Leal, Kuang-Huei Lee, Sergey Levine, Yao Lu, Utsav Malla, Deeksha Manjunath, Igor Mordatch, Ofir Nachum, Carolina Parada, Jodilyn Peralta, Emily Perez, Karl Pertsch, Jornell Quiambao, Kanishka Rao, Michael S. Ryoo, Grecia Salazar, Pannag Sanketi, Kevin Sayed, Jaspiar Singh, Sumedh Sontakke, Austin Stone, Clayton Tan, Huong T. Tran, Vincent Vanhoucke
, Steve Vega, Quan Vuong
, Fei Xia, Ted Xiao, Peng Xu, Sichun Xu, Tianhe Yu, Brianna Zitkovich:
RT-1: Robotics Transformer for Real-World Control at Scale. CoRR abs/2212.06817 (2022) - [i321]Dhruv Shah, Arjun Bhorkar, Hrish Leen, Ilya Kostrikov, Nick Rhinehart, Sergey Levine:
Offline Reinforcement Learning for Visual Navigation. CoRR abs/2212.08244 (2022) - [i320]Kelvin Xu, Zheyuan Hu, Ria Doshi, Aaron Rovinsky, Vikash Kumar, Abhishek Gupta, Sergey Levine:
Dexterous Manipulation from Images: Autonomous Real-World RL via Substep Guidance. CoRR abs/2212.09902 (2022) - [i319]Yiren Lu, Justin Fu, George Tucker, Xinlei Pan, Eli Bronstein, Becca Roelofs, Benjamin Sapp, Brandyn White, Aleksandra Faust
, Shimon Whiteson, Dragomir Anguelov, Sergey Levine:
Imitation Is Not Enough: Robustifying Imitation with Reinforcement Learning for Challenging Driving Scenarios. CoRR abs/2212.11419 (2022) - [i318]Qiyang Li, Yuexiang Zhai, Yi Ma, Sergey Levine:
Understanding the Complexity Gains of Single-Task RL with a Curriculum. CoRR abs/2212.12809 (2022) - 2021
- [j19]Julian Ibarz
, Jie Tan, Chelsea Finn
, Mrinal Kalakrishnan
, Peter Pastor, Sergey Levine:
How to train your robot with deep reinforcement learning: lessons we have learned. Int. J. Robotics Res. 40(4-5) (2021) - [j18]Gregory Kahn
, Pieter Abbeel, Sergey Levine:
BADGR: An Autonomous Self-Supervised Learning-Based Navigation System. IEEE Robotics Autom. Lett. 6(2): 1312-1319 (2021) - [j17]Suneel Belkhale
, Rachel Li, Gregory Kahn
, Rowan McAllister
, Roberto Calandra
, Sergey Levine:
Model-Based Meta-Reinforcement Learning for Flight With Suspended Payloads. IEEE Robotics Autom. Lett. 6(2): 1471-1478 (2021) - [j16]Gregory Kahn
, Pieter Abbeel, Sergey Levine
:
LaND: Learning to Navigate From Disengagements. IEEE Robotics Autom. Lett. 6(2): 1872-1879 (2021) - [j15]Xue Bin Peng, Ze Ma, Pieter Abbeel, Sergey Levine, Angjoo Kanazawa
:
AMP: adversarial motion priors for stylized physics-based character control. ACM Trans. Graph. 40(4): 144:1-144:20 (2021) - [c276]Charles Sun, Jedrzej Orbik, Coline Manon Devin, Brian H. Yang, Abhishek Gupta, Glen Berseth, Sergey Levine:
Fully Autonomous Real-World Reinforcement Learning with Applications to Mobile Manipulation. CoRL 2021: 308-319 - [c275]Aviral Kumar, Anikait Singh, Stephen Tian, Chelsea Finn, Sergey Levine:
A Workflow for Offline Model-Free Robotic Reinforcement Learning. CoRL 2021: 417-428 - [c274]Dmitry Kalashnikov, Jake Varley, Yevgen Chebotar, Benjamin Swanson, Rico Jonschkowski, Chelsea Finn, Sergey Levine, Karol Hausman:
Scaling Up Multi-Task Robotic Reinforcement Learning. CoRL 2021: 557-575 - [c273]Dhruv Shah, Benjamin Eysenbach, Nicholas Rhinehart, Sergey Levine:
Rapid Exploration for Open-World Navigation with Latent Goal Models. CoRL 2021: 674-684 - [c272]Eric Jang, Alex Irpan, Mohi Khansari, Daniel Kappler, Frederik Ebert, Corey Lynch, Sergey Levine, Chelsea Finn:
BC-Z: Zero-Shot Task Generalization with Robotic Imitation Learning. CoRL 2021: 991-1002 - [c271]Yao Lu, Karol Hausman, Yevgen Chebotar, Mengyuan Yan, Eric Jang, Alexander Herzog, Ted Xiao, Alex Irpan, Mohi Khansari, Dmitry Kalashnikov, Sergey Levine:
AW-Opt: Learning Robotic Skills with Imitation andReinforcement at Scale. CoRL 2021: 1078-1088 - [c270]Katie Kang, Gregory Kahn, Sergey Levine:
Hierarchically Integrated Models: Learning to Navigate from Heterogeneous Robots. CoRL 2021: 1316-1325 - [c269]Sergey Levine:
Understanding the World Through Action. CoRL 2021: 1752-1757 - [c268]Amy Zhang, Rowan Thomas McAllister, Roberto Calandra, Yarin Gal, Sergey Levine:
Learning Invariant Representations for Reinforcement Learning without Reconstruction. ICLR 2021 - [c267]Anurag Ajay, Aviral Kumar, Pulkit Agrawal, Sergey Levine, Ofir Nachum:
OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning. ICLR 2021 - [c266]Glen Berseth
, Daniel Geng, Coline Manon Devin, Nicholas Rhinehart, Chelsea Finn, Dinesh Jayaraman, Sergey Levine:
SMiRL: Surprise Minimizing Reinforcement Learning in Unstable Environments. ICLR 2021 - [c265]Homanga Bharadhwaj, Aviral Kumar, Nicholas Rhinehart, Sergey Levine, Florian Shkurti, Animesh Garg:
Conservative Safety Critics for Exploration. ICLR 2021 - [c264]John D. Co-Reyes, Yingjie Miao, Daiyi Peng, Esteban Real, Quoc V. Le, Sergey Levine, Honglak Lee, Aleksandra Faust:
Evolving Reinforcement Learning Algorithms. ICLR 2021 - [c263]Benjamin Eysenbach, Shreyas Chaudhari, Swapnil Asawa, Sergey Levine, Ruslan Salakhutdinov:
Off-Dynamics Reinforcement Learning: Training for Transfer with Domain Classifiers. ICLR 2021 - [c262]Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine:
C-Learning: Learning to Achieve Goals via Recursive Classification. ICLR 2021 - [c261]Justin Fu, Mohammad Norouzi, Ofir Nachum, George Tucker, Ziyu Wang, Alexander Novikov, Mengjiao Yang, Michael R. Zhang, Yutian Chen, Aviral Kumar, Cosmin Paduraru, Sergey Levine, Tom Le Paine:
Benchmarks for Deep Off-Policy Evaluation. ICLR 2021 - [c260]Justin Fu, Sergey Levine:
Offline Model-Based Optimization via Normalized Maximum Likelihood Estimation. ICLR 2021 - [c259]Jensen Gao, Siddharth Reddy, Glen Berseth, Nicholas Hardy, Nikhilesh Natraj, Karunesh Ganguly, Anca D. Dragan, Sergey Levine:
X2T: Training an X-to-Text Typing Interface with Online Learning from User Feedback. ICLR 2021 - [c258]Dibya Ghosh, Abhishek Gupta, Ashwin Reddy, Justin Fu, Coline Manon Devin, Benjamin Eysenbach, Sergey Levine:
Learning to Reach Goals via Iterated Supervised Learning. ICLR 2021 - [c257]Anirudh Goyal, Alex Lamb, Phanideep Gampa, Philippe Beaudoin, Charles Blundell, Sergey Levine, Yoshua Bengio, Michael Curtis Mozer:
Factorizing Declarative and Procedural Knowledge in Structured, Dynamical Environments. ICLR 2021 - [c256]Anirudh Goyal, Alex Lamb, Jordan Hoffmann, Shagun Sodhani, Sergey Levine, Yoshua Bengio, Bernhard Schölkopf:
Recurrent Independent Mechanisms. ICLR 2021 - [c255]Aviral Kumar, Rishabh Agarwal, Dibya Ghosh, Sergey Levine:
Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning. ICLR 2021 - [c254]Avi Singh, Huihan Liu, Gaoyue Zhou, Albert Yu, Nicholas Rhinehart, Sergey Levine:
Parrot: Data-Driven Behavioral Priors for Reinforcement Learning. ICLR 2021 - [c253]Stephen Tian, Suraj Nair, Frederik Ebert, Sudeep Dasari, Benjamin Eysenbach, Chelsea Finn, Sergey Levine:
Model-Based Visual Planning with Self-Supervised Functional Distances. ICLR 2021 - [c252]Michael Chang, Sidhant Kaushik, Sergey Levine, Tom Griffiths:
Modularity in Reinforcement Learning via Algorithmic Independence in Credit Assignment. ICML 2021: 1452-1462 - [c251]Yevgen Chebotar, Karol Hausman, Yao Lu, Ted Xiao, Dmitry Kalashnikov, Jacob Varley, Alex Irpan, Benjamin Eysenbach, Ryan Julian, Chelsea Finn, Sergey Levine:
Actionable Models: Unsupervised Offline Reinforcement Learning of Robotic Skills. ICML 2021: 1518-1528 - [c250]Jongwook Choi, Archit Sharma, Honglak Lee, Sergey Levine, Shixiang Shane Gu:
Variational Empowerment as Representation Learning for Goal-Conditioned Reinforcement Learning. ICML 2021: 1953-1963 - [c249]Angelos Filos, Clare Lyle, Yarin Gal, Sergey Levine, Natasha Jaques, Gregory Farquhar:
PsiPhi-Learning: Reinforcement Learning with Demonstrations using Successor Features and Inverse Temporal Difference Learning. ICML 2021: 3305-3317 - [c248]Hiroki Furuta, Tatsuya Matsushima, Tadashi Kozuno, Yutaka Matsuo, Sergey Levine, Ofir Nachum, Shixiang Shane Gu:
Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning. ICML 2021: 3541-3552 - [c247]Pang Wei Koh, Shiori Sagawa, Henrik Marklund, Sang Michael Xie, Marvin Zhang, Akshay Balsubramani, Weihua Hu, Michihiro Yasunaga, Richard Lanas Phillips, Irena Gao, Tony Lee, Etienne David, Ian Stavness, Wei Guo, Berton Earnshaw, Imran S. Haque, Sara M. Beery, Jure Leskovec, Anshul Kundaje, Emma Pierson, Sergey Levine, Chelsea Finn, Percy Liang:
WILDS: A Benchmark of in-the-Wild Distribution Shifts. ICML 2021: 5637-5664 - [c246]Kevin Li, Abhishek Gupta, Ashwin Reddy, Vitchyr H. Pong, Aurick Zhou, Justin Yu, Sergey Levine:
MURAL: Meta-Learning Uncertainty-Aware Rewards for Outcome-Driven Reinforcement Learning. ICML 2021: 6346-6356 - [c245]Eric Mitchell, Rafael Rafailov, Xue Bin Peng, Sergey Levine, Chelsea Finn:
Offline Meta-Reinforcement Learning with Advantage Weighting. ICML 2021: 7780-7791 - [c244]Kamal Ndousse, Douglas Eck, Sergey Levine, Natasha Jaques:
Emergent Social Learning via Multi-agent Reinforcement Learning. ICML 2021: 7991-8004 - [c243]Oleh Rybkin, Kostas Daniilidis, Sergey Levine:
Simple and Effective VAE Training with Calibrated Decoders. ICML 2021: 9179-9189 - [c242]Oleh Rybkin, Chuning Zhu, Anusha Nagabandi, Kostas Daniilidis, Igor Mordatch, Sergey Levine:
Model-Based Reinforcement Learning via Latent-Space Collocation. ICML 2021: 9190-9201 - [c241]Brandon Trabucco, Aviral Kumar, Xinyang Geng, Sergey Levine:
Conservative Objective Models for Effective Offline Model-Based Optimization. ICML 2021: 10358-10368 - [c240]Aurick Zhou, Sergey Levine:
Amortized Conditional Normalized Maximum Likelihood: Reliable Out of Distribution Uncertainty Estimation. ICML 2021: 12803-12812 - [c239]Zhongyu Li, Xuxin Cheng, Xue Bin Peng, Pieter Abbeel, Sergey Levine, Glen Berseth, Koushil Sreenath:
Reinforcement Learning for Robust Parameterized Locomotion Control of Bipedal Robots. ICRA 2021: 2811-2817 - [c238]Yifeng Jiang, Tingnan Zhang, Daniel Ho, Yunfei Bai, C. Karen Liu, Sergey Levine, Jie Tan:
SimGAN: Hybrid Simulator Identification for Domain Adaptation via Adversarial Reinforcement Learning. ICRA 2021: 2884-2890 - [c237]Soroush Nasiriany, Vitchyr H. Pong, Ashvin Nair, Alexander Khazatsky, Glen Berseth, Sergey Levine:
DisCo RL: Distribution-Conditioned Reinforcement Learning for General-Purpose Policies. ICRA 2021: 6635-6641 - [c236]Abhishek Gupta, Justin Yu, Tony Z. Zhao, Vikash Kumar, Aaron Rovinsky, Kelvin Xu, Thomas Devlin, Sergey Levine:
Reset-Free Reinforcement Learning via Multi-Task Learning: Learning Dexterous Manipulation Behaviors without Human Intervention. ICRA 2021: 6664-6671 - [c235]Dhruv Shah, Benjamin Eysenbach
, Gregory Kahn, Nicholas Rhinehart
, Sergey Levine:
ViNG: Learning Open-World Navigation with Visual Goals. ICRA 2021: 13215-13222 - [c234]Nicholas Rhinehart
, Jeff He, Charles Packer, Matthew A. Wright, Rowan McAllister, Joseph E. Gonzalez
, Sergey Levine:
Contingencies from Observations: Tractable Contingency Planning with Learned Behavior Models. ICRA 2021: 13663-13669 - [c233]Alexander Khazatsky, Ashvin Nair, Daniel Jing, Sergey Levine:
What Can I Do Here? Learning New Skills by Imagining Visual Affordances. ICRA 2021: 14291-14297 - [c232]Aurick Zhou, Sergey Levine:
Bayesian Adaptation for Covariate Shift. NeurIPS 2021: 914-927 - [c231]Michael Janner, Qiyang Li, Sergey Levine:
Offline Reinforcement Learning as One Big Sequence Modeling Problem. NeurIPS 2021: 1273-1286 - [c230]Nicholas Rhinehart, Jenny Wang, Glen Berseth, John D. Co-Reyes, Danijar Hafner, Chelsea Finn, Sergey Levine:
Information is Power: Intrinsic Control via Information Capture. NeurIPS 2021: 10745-10758 - [c229]Tianhe Yu, Aviral Kumar, Yevgen Chebotar, Karol Hausman, Sergey Levine, Chelsea Finn:
Conservative Data Sharing for Multi-Task Offline Reinforcement Learning. NeurIPS 2021: 11501-11516 - [c228]Ben Eysenbach, Sergey Levine, Ruslan Salakhutdinov:
Replacing Rewards with Examples: Example-Based Policy Search via Recursive Classification. NeurIPS 2021: 11541-11552 - [c227]Tim G. J. Rudner, Vitchyr Pong, Rowan McAllister, Yarin Gal, Sergey Levine:
Outcome-Driven Reinforcement Learning via Variational Inference. NeurIPS 2021: 13045-13058 - [c226]Archit Sharma, Abhishek Gupta, Sergey Levine, Karol Hausman, Chelsea Finn:
Autonomous Reinforcement Learning via Subgoal Curricula. NeurIPS 2021: 18474-18486 - [c225]Marvin Zhang, Henrik Marklund, Nikita Dhawan, Abhishek Gupta, Sergey Levine, Chelsea Finn:
Adaptive Risk Minimization: Learning to Adapt to Domain Shift. NeurIPS 2021: 23664-23678 - [c224]Dibya Ghosh, Jad Rahme, Aviral Kumar, Amy Zhang, Ryan P. Adams, Sergey Levine:
Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability. NeurIPS 2021: 25502-25515 - [c223]Kate Rakelly, Abhishek Gupta, Carlos Florensa, Sergey Levine:
Which Mutual-Information Representation Learning Objectives are Sufficient for Control? NeurIPS 2021: 26345-26357 - [c222]Siddharth Reddy, Anca D. Dragan, Sergey Levine:
Pragmatic Image Compression for Human-in-the-Loop Decision-Making. NeurIPS 2021: 26499-26510 - [c221]Ben Eysenbach, Ruslan Salakhutdinov, Sergey Levine:
Robust Predictable Control. NeurIPS 2021: 27813-27825 - [c220]Tianhe Yu, Aviral Kumar, Rafael Rafailov, Aravind Rajeswaran, Sergey Levine, Chelsea Finn:
COMBO: Conservative Offline Model-Based Policy Optimization. NeurIPS 2021: 28954-28967 - [i317]John D. Co-Reyes, Yingjie Miao, Daiyi Peng, Esteban Real, Sergey Levine, Quoc V. Le, Honglak Lee, Aleksandra Faust:
Evolving Reinforcement Learning Algorithms. CoRR abs/2101.03958 (2021) - [i316]Yifeng Jiang, Tingnan Zhang, Daniel Ho, Yunfei Bai, C. Karen Liu, Sergey Levine, Jie Tan:
SimGAN: Hybrid Simulator Identification for Domain Adaptation via Adversarial Reinforcement Learning. CoRR abs/2101.06005 (2021) - [i315]Julian Ibarz, Jie Tan, Chelsea Finn, Mrinal Kalakrishnan, Peter Pastor, Sergey Levine:
How to Train Your Robot with Deep Reinforcement Learning; Lessons We've Learned. CoRR abs/2102.02915 (2021) - [i314]Justin Fu, Sergey Levine:
Offline Model-Based Optimization via Normalized Maximum Likelihood Estimation. CoRR abs/2102.07970 (2021) - [i313]Tianhe Yu, Aviral Kumar, Rafael Rafailov, Aravind Rajeswaran, Sergey Levine, Chelsea Finn:
COMBO: Conservative Offline Model-Based Policy Optimization. CoRR abs/2102.08363 (2021) - [i312]Angelos Filos, Clare Lyle, Yarin Gal, Sergey Levine, Natasha Jaques, Gregory Farquhar:
PsiPhi-Learning: Reinforcement Learning with Demonstrations using Successor Features and Inverse Temporal Difference Learning. CoRR abs/2102.12560 (2021) - [i311]Benjamin Eysenbach, Sergey Levine:
Maximum Entropy RL (Provably) Solves Some Robust RL Problems. CoRR abs/2103.06257 (2021) - [i310]Benjamin Eysenbach, Sergey Levine, Ruslan Salakhutdinov:
Replacing Rewards with Examples: Example-Based Policy Search via Recursive Classification. CoRR abs/2103.12656 (2021) - [i309]Hiroki Furuta, Tatsuya Matsushima, Tadashi Kozuno, Yutaka Matsuo, Sergey Levine, Ofir Nachum, Shixiang Shane Gu:
Policy Information Capacity: Information-Theoretic Measure for Task Complexity in Deep Reinforcement Learning. CoRR abs/2103.12726 (2021) - [i308]Zhongyu Li, Xuxin Cheng, Xue Bin Peng, Pieter Abbeel, Sergey Levine, Glen Berseth, Koushil Sreenath:
Reinforcement Learning for Robust Parameterized Locomotion Control of Bipedal Robots. CoRR abs/2103.14295 (2021) - [i307]Justin Fu, Mohammad Norouzi, Ofir Nachum, George Tucker, Ziyu Wang, Alexander Novikov, Mengjiao Yang, Michael R. Zhang, Yutian Chen, Aviral Kumar, Cosmin Paduraru, Sergey Levine, Tom Le Paine:
Benchmarks for Deep Off-Policy Evaluation. CoRR abs/2103.16596 (2021) - [i306]Xue Bin Peng, Ze Ma, Pieter Abbeel, Sergey Levine, Angjoo Kanazawa:
AMP: Adversarial Motion Priors for Stylized Physics-Based Character Control. CoRR abs/2104.02180 (2021) - [i305]Dhruv Shah, Benjamin Eysenbach, Nicholas Rhinehart, Sergey Levine:
RECON: Rapid Exploration for Open-World Navigation with Latent Goal Models. CoRR abs/2104.05859 (2021) - [i304]Yevgen Chebotar, Karol Hausman, Yao Lu, Ted Xiao, Dmitry Kalashnikov, Jake Varley, Alex Irpan, Benjamin Eysenbach, Ryan Julian, Chelsea Finn, Sergey Levine:
Actionable Models: Unsupervised Offline Reinforcement Learning of Robotic Skills. CoRR abs/2104.07749 (2021) - [i303]Dmitry Kalashnikov, Jacob Varley, Yevgen Chebotar, Benjamin Swanson, Rico Jonschkowski, Chelsea Finn, Sergey Levine, Karol Hausman:
MT-Opt: Continuous Multi-Task Robotic Reinforcement Learning at Scale. CoRR abs/2104.08212 (2021) - [i302]Tim G. J. Rudner
, Vitchyr H. Pong, Rowan McAllister, Yarin Gal, Sergey Levine:
Outcome-Driven Reinforcement Learning via Variational Inference. CoRR abs/2104.10190 (2021) - [i301]Nicholas Rhinehart, Jeff He, Charles Packer, Matthew A. Wright, Rowan McAllister, Joseph E. Gonzalez, Sergey Levine:
Contingencies from Observations: Tractable Contingency Planning with Learned Behavior Models. CoRR abs/2104.10558 (2021) - [i300]Abhishek Gupta, Justin Yu, Tony Z. Zhao, Vikash Kumar, Aaron Rovinsky, Kelvin Xu, Thomas Devlin, Sergey Levine:
Reset-Free Reinforcement Learning via Multi-Task Learning: Learning Dexterous Manipulation Behaviors without Human Intervention. CoRR abs/2104.11203 (2021) - [i299]Soroush Nasiriany, Vitchyr H. Pong, Ashvin Nair, Alexander Khazatsky, Glen Berseth, Sergey Levine:
DisCo RL: Distribution-Conditioned Reinforcement Learning for General-Purpose Policies. CoRR abs/2104.11707 (2021) - [i298]Alexander Khazatsky, Ashvin Nair, Daniel Jing, Sergey Levine:
What Can I Do Here? Learning New Skills by Imagining Visual Affordances. CoRR abs/2106.00671 (2021) - [i297]Jongwook Choi, Archit Sharma, Honglak Lee, Sergey Levine, Shixiang Shane Gu:
Variational Empowerment as Representation Learning for Goal-Based Reinforcement Learning. CoRR abs/2106.01404 (2021) - [i296]Michael Janner, Qiyang Li, Sergey Levine:
Reinforcement Learning as One Big Sequence Modeling Problem. CoRR abs/2106.02039 (2021) - [i295]Kate Rakelly, Abhishek Gupta, Carlos Florensa, Sergey Levine:
Which Mutual-Information Representation Learning Objectives are Sufficient for Control? CoRR abs/2106.07278 (2021) - [i294]Mohammad Babaeizadeh, Mohammad Taghi Saffar, Suraj Nair, Sergey Levine, Chelsea Finn, Dumitru Erhan:
FitVid: Overfitting in Pixel-Level Video Prediction. CoRR abs/2106.13195 (2021) - [i293]Oleh Rybkin, Chuning Zhu, Anusha Nagabandi, Kostas Daniilidis, Igor Mordatch, Sergey Levine:
Model-Based Reinforcement Learning via Latent-Space Collocation. CoRR abs/2106.13229 (2021) - [i292]Katie Kang, Gregory Kahn, Sergey Levine:
Multi-Robot Deep Reinforcement Learning for Mobile Navigation. CoRR abs/2106.13280 (2021) - [i291]Michael Chang, Sidhant Kaushik, Sergey Levine, Thomas L. Griffiths:
Modularity in Reinforcement Learning via Algorithmic Independence in Credit Assignment. CoRR abs/2106.14993 (2021) - [i290]Vitchyr H. Pong, Ashvin Nair, Laura M. Smith, Catherine Huang, Sergey Levine:
Offline Meta-Reinforcement Learning with Online Self-Supervision. CoRR abs/2107.03974 (2021) - [i289]Dibya Ghosh, Jad Rahme, Aviral Kumar, Amy Zhang, Ryan P. Adams, Sergey Levine:
Why Generalization in RL is Difficult: Epistemic POMDPs and Implicit Partial Observability. CoRR abs/2107.06277 (2021) - [i288]Brandon Trabucco, Aviral Kumar, Xinyang Geng, Sergey Levine:
Conservative Objective Models for Effective Offline Model-Based Optimization. CoRR abs/2107.06882 (2021) - [i287]Kevin Li, Abhishek Gupta, Ashwin Reddy, Vitchyr Pong, Aurick Zhou, Justin Yu, Sergey Levine:
MURAL: Meta-Learning Uncertainty-Aware Rewards for Outcome-Driven Reinforcement Learning. CoRR abs/2107.07184 (2021) - [i286]Arnaud Fickinger, Natasha Jaques
, Samyak Parajuli, Michael Chang, Nicholas Rhinehart, Glen Berseth, Stuart Russell, Sergey Levine:
Explore and Control with Adversarial Surprise. CoRR abs/2107.07394 (2021) - [i285]Archit Sharma, Abhishek Gupta, Sergey Levine, Karol Hausman, Chelsea Finn:
Persistent Reinforcement Learning via Subgoal Curricula. CoRR abs/2107.12931 (2021) - [i284]Charles Sun, Jedrzej Orbik, Coline Devin, Brian H. Yang, Abhishek Gupta, Glen Berseth, Sergey Levine:
ReLMM: Practical RL for Learning Mobile Manipulation Skills Using Only Onboard Sensors. CoRR abs/2107.13545 (2021) - [i283]Siddharth Reddy, Anca D. Dragan, Sergey Levine:
Pragmatic Image Compression for Human-in-the-Loop Decision-Making. CoRR abs/2108.04219 (2021) - [i282]Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine:
Robust Predictable Control. CoRR abs/2109.03214 (2021) - [i281]Tianhe Yu, Aviral Kumar, Yevgen Chebotar, Karol Hausman, Sergey Levine, Chelsea Finn:
Conservative Data Sharing for Multi-Task Offline Reinforcement Learning. CoRR abs/2109.08128 (2021) - [i280]Aviral Kumar, Anikait Singh, Stephen Tian, Chelsea Finn, Sergey Levine:
A Workflow for Offline Model-Free Robotic Reinforcement Learning. CoRR abs/2109.10813 (2021) - [i279]Aurick Zhou, Sergey Levine:
Training on Test Data with Bayesian Adaptation for Covariate Shift. CoRR abs/2109.12746 (2021) - [i278]Frederik Ebert, Yanlai Yang, Karl Schmeckpeper, Bernadette Bucher, Georgios Georgakis, Kostas Daniilidis, Chelsea Finn, Sergey Levine:
Bridge Data: Boosting Generalization of Robotic Skills with Cross-Domain Datasets. CoRR abs/2109.13396 (2021) - [i277]Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine:
The Information Geometry of Unsupervised Reinforcement Learning. CoRR abs/2110.02719 (2021) - [i276]Benjamin Eysenbach, Alexander Khazatsky, Sergey Levine, Ruslan Salakhutdinov:
Mismatched No More: Joint Model-Policy Optimization for Model-Based RL. CoRR abs/2110.02758 (2021) - [i275]Tony Z. Zhao, Jianlan Luo, Oleg Sushkov, Rugile Pevceviciute, Nicolas Heess, Jonathan Scholz, Stefan Schaal, Sergey Levine:
Offline Meta-Reinforcement Learning for Industrial Insertion. CoRR abs/2110.04276 (2021) - [i274]Laura M. Smith, J. Chase Kew, Xue Bin Peng, Sehoon Ha, Jie Tan, Sergey Levine:
Legged Robots that Keep on Learning: Fine-Tuning Locomotion Policies in the Real World. CoRR abs/2110.05457 (2021) - [i273]Ilya Kostrikov, Ashvin Nair, Sergey Levine:
Offline Reinforcement Learning with Implicit Q-Learning. CoRR abs/2110.06169 (2021) - [i272]Marvin Zhang, Sergey Levine, Chelsea Finn:
MEMO: Test Time Robustness via Adaptation and Augmentation. CoRR abs/2110.09506 (2021) - [i271]Aviral Kumar, Amir Yazdanbakhsh, Milad Hashemi, Kevin Swersky, Sergey Levine:
Data-Driven Offline Optimization For Architecting Hardware Accelerators. CoRR abs/2110.11346 (2021) - [i270]Tianjun Zhang, Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine, Joseph E. Gonzalez:
C-Planning: An Automatic Curriculum for Learning Goal-Reaching Tasks. CoRR abs/2110.12080 (2021) - [i269]Sergey Levine:
Understanding the World Through Action. CoRR abs/2110.12543 (2021) - [i268]Mengjiao Yang, Sergey Levine, Ofir Nachum:
TRAIL: Near-Optimal Imitation Learning with Suboptimal Data. CoRR abs/2110.14770 (2021) - [i267]Dhruv Shah, Peng Xu, Yao Lu, Ted Xiao, Alexander Toshev, Sergey Levine, Brian Ichter:
Value Function Spaces: Skill-Centric State Abstractions for Long-Horizon Reasoning. CoRR abs/2111.03189 (2021) - [i266]Yao Lu, Karol Hausman, Yevgen Chebotar, Mengyuan Yan, Eric Jang, Alexander Herzog, Ted Xiao, Alex Irpan, Mohi Khansari, Dmitry Kalashnikov, Sergey Levine:
AW-Opt: Learning Robotic Skills with Imitation and Reinforcement at Scale. CoRR abs/2111.05424 (2021) - [i265]Nitish Dashora, Daniel Shin, Dhruv Shah, Henry A. Leopold, David D. Fan, Ali-Akbar Agha-Mohammadi, Nicholas Rhinehart, Sergey Levine:
Hybrid Imitative Planning with Geometric and Predictive Costs in Off-road Environments. CoRR abs/2111.10948 (2021) - [i264]Nicholas Rhinehart, Jenny Wang, Glen Berseth, John D. Co-Reyes, Danijar Hafner, Chelsea Finn, Sergey Levine:
Information is Power: Intrinsic Control via Information Capture. CoRR abs/2112.03899 (2021) - [i263]Glen Berseth, Zhiwei Zhang, Grace Zhang, Chelsea Finn, Sergey Levine:
CoMPS: Continual Meta Policy Search. CoRR abs/2112.04467 (2021) - [i262]Aviral Kumar, Rishabh Agarwal, Tengyu Ma, Aaron C. Courville, George Tucker, Sergey Levine:
DR3: Value-Based Deep Reinforcement Learning Requires Explicit Regularization. CoRR abs/2112.04716 (2021) - [i261]Shiori Sagawa, Pang Wei Koh, Tony Lee, Irena Gao, Sang Michael Xie, Kendrick Shen, Ananya Kumar, Weihua Hu, Michihiro Yasunaga, Henrik Marklund, Sara Beery, Etienne David, Ian Stavness, Wei Guo, Jure Leskovec, Kate Saenko, Tatsunori Hashimoto, Sergey Levine, Chelsea Finn, Percy Liang:
Extending the WILDS Benchmark for Unsupervised Adaptation. CoRR abs/2112.05090 (2021) - [i260]Archit Sharma, Kelvin Xu, Nikhil Sardana, Abhishek Gupta, Karol Hausman, Sergey Levine, Chelsea Finn:
Autonomous Reinforcement Learning: Formalism and Benchmarking. CoRR abs/2112.09605 (2021) - [i259]Scott Emmons, Benjamin Eysenbach, Ilya Kostrikov, Sergey Levine:
RvS: What is Essential for Offline RL via Supervised Learning? CoRR abs/2112.10751 (2021) - 2020
- [j14]Saurabh Gupta, Varun Tolani, James Davidson, Sergey Levine, Rahul Sukthankar, Jitendra Malik:
Cognitive Mapping and Planning for Visual Navigation. Int. J. Comput. Vis. 128(5): 1311-1330 (2020) - [j13]Brian H. Yang
, Dinesh Jayaraman
, Glen Berseth
, Alexei A. Efros
, Sergey Levine:
Morphology-Agnostic Visual Robotic Control. IEEE Robotics Autom. Lett. 5(2): 766-773 (2020) - [j12]Brijen Thananjeyan
, Ashwin Balakrishna
, Ugo Rosolia
, Felix Li, Rowan McAllister, Joseph E. Gonzalez
, Sergey Levine, Francesco Borrelli
, Ken Goldberg
:
Safety Augmented Value Estimation From Demonstrations (SAVED): Safe Deep Model-Based RL for Sparse Cost Robotic Tasks. IEEE Robotics Autom. Lett. 5(2): 3612-3619 (2020) - [c219]Sergey Levine:
Unsupervised Reinforcement Learning. AAMAS 2020: 5-6 - [c218]Xinshuo Weng, Jianren Wang, Sergey Levine, Kris Kitani, Nicholas Rhinehart:
Inverting the Pose Forecasting Pipeline with SPF2: Sequential Pointcloud Forecasting for Sequential Pose Forecasting. CoRL 2020: 11-20 - [c217]Karl Schmeckpeper, Oleh Rybkin, Kostas Daniilidis, Sergey Levine, Chelsea Finn:
Reinforcement Learning with Videos: Combining Offline Observations with Interaction. CoRL 2020: 339-354 - [c216]Siddharth Reddy, Sergey Levine, Anca D. Dragan:
Assisted Perception: Optimizing Observations to Communicate State. CoRL 2020: 748-764 - [c215]Sehoon Ha, Peng Xu, Zhenyu Tan, Sergey Levine, Jie Tan:
Learning to Walk in the Real World with Minimal Human Effort. CoRL 2020: 1110-1120 - [c214]Zihao Zhao, Anusha Nagabandi, Kate Rakelly, Chelsea Finn, Sergey Levine:
MELD: Meta-Reinforcement Learning from Images via Latent State Models. CoRL 2020: 1246-1261 - [c213]Ryan Julian, Benjamin Swanson, Gaurav S. Sukhatme, Sergey Levine, Chelsea Finn, Karol Hausman:
Never Stop Learning: The Effectiveness of Fine-Tuning in Robotic Reinforcement Learning. CoRL 2020: 2120-2136 - [c212]Avi Singh, Albert Yu, Jonathan Yang, Jesse Zhang, Aviral Kumar, Sergey Levine:
Chaining Behaviors from Data with Model-Free Reinforcement Learning. CoRL 2020: 2162-2177 - [c211]Kanishka Rao, Chris Harris, Alex Irpan, Sergey Levine, Julian Ibarz, Mohi Khansari:
RL-CycleGAN: Reinforcement Learning Aware Simulation-to-Real. CVPR 2020: 11154-11163 - [c210]Karl Schmeckpeper
, Annie Xie
, Oleh Rybkin
, Stephen Tian
, Kostas Daniilidis
, Sergey Levine
, Chelsea Finn
:
Learning Predictive Models from Observation and Interaction. ECCV (20) 2020: 708-725 - [c209]Adam Gleave, Michael Dennis, Cody Wild, Neel Kant, Sergey Levine, Stuart Russell:
Adversarial Policies: Attacking Deep Reinforcement Learning. ICLR 2020 - [c208]Anirudh Goyal, Yoshua Bengio, Matthew M. Botvinick, Sergey Levine:
The Variational Bandwidth Bottleneck: Stochastic Evaluation on an Information Budget. ICLR 2020 - [c207]Anirudh Goyal, Shagun Sodhani, Jonathan Binas, Xue Bin Peng, Sergey Levine, Yoshua Bengio:
Reinforcement Learning with Competitive Ensembles of Information-Constrained Primitives. ICLR 2020 - [c206]Kristian Hartikainen, Xinyang Geng, Tuomas Haarnoja, Sergey Levine:
Dynamical Distance Learning for Semi-Supervised and Unsupervised Skill Discovery. ICLR 2020 - [c205]Lukasz Kaiser, Mohammad Babaeizadeh, Piotr Milos, Blazej Osinski, Roy H. Campbell, Konrad Czechowski, Dumitru Erhan, Chelsea Finn, Piotr Kozakowski
, Sergey Levine, Afroz Mohiuddin, Ryan Sepassi, George Tucker, Henryk Michalewski:
Model Based Reinforcement Learning for Atari. ICLR 2020 - [c204]Manoj Kumar, Mohammad Babaeizadeh, Dumitru Erhan, Chelsea Finn, Sergey Levine, Laurent Dinh, Durk Kingma:
VideoFlow: A Conditional Flow-Based Model for Stochastic Video Generation. ICLR 2020 - [c203]Siddharth Reddy, Anca D. Dragan, Sergey Levine:
SQIL: Imitation Learning via Reinforcement Learning with Sparse Rewards. ICLR 2020 - [c202]Nicholas Rhinehart, Rowan McAllister, Sergey Levine:
Deep Imitative Models for Flexible Inference, Planning, and Control. ICLR 2020 - [c201]Archit Sharma, Shixiang Gu, Sergey Levine, Vikash Kumar, Karol Hausman:
Dynamics-Aware Unsupervised Discovery of Skills. ICLR 2020 - [c200]Ted Xiao, Eric Jang, Dmitry Kalashnikov, Sergey Levine, Julian Ibarz, Karol Hausman, Alexander Herzog:
Thinking While Moving: Deep Reinforcement Learning with Concurrent Control. ICLR 2020 - [c199]Mingzhang Yin, George Tucker, Mingyuan Zhou
, Sergey Levine, Chelsea Finn:
Meta-Learning without Memorization. ICLR 2020 - [c198]Allan Zhou, Eric Jang, Daniel Kappler, Alexander Herzog, Mohi Khansari, Paul Wohlhart, Yunfei Bai, Mrinal Kalakrishnan, Sergey Levine, Chelsea Finn:
Watch, Try, Learn: Meta-Learning from Demonstrations and Rewards. ICLR 2020 - [c197]Henry Zhu, Justin Yu, Abhishek Gupta, Dhruv Shah, Kristian Hartikainen, Avi Singh, Vikash Kumar, Sergey Levine:
The Ingredients of Real World Robotic Reinforcement Learning. ICLR 2020 - [c196]Michael Chang, Sidhant Kaushik, S. Matthew Weinberg
, Tom Griffiths, Sergey Levine:
Decentralized Reinforcement Learning: Global Decision-Making via Local Economic Transactions. ICML 2020: 1437-1447 - [c195]Angelos Filos, Panagiotis Tigas, Rowan McAllister, Nicholas Rhinehart, Sergey Levine, Yarin Gal:
Can Autonomous Vehicles Identify, Recover From, and Adapt to Distribution Shifts? ICML 2020: 3145-3153 - [c194]Vitchyr Pong, Murtaza Dalal, Steven Lin, Ashvin Nair, Shikhar Bahl, Sergey Levine:
Skew-Fit: State-Covering Self-Supervised Reinforcement Learning. ICML 2020: 7783-7792 - [c193]Siddharth Reddy, Anca D. Dragan, Sergey Levine, Shane Legg, Jan Leike:
Learning Human Objectives by Evaluating Hypothetical Behavior. ICML 2020: 8020-8029 - [c192]Jesse Zhang, Brian Cheung, Chelsea Finn, Sergey Levine, Dinesh Jayaraman:
Cautious Adaptation For Reinforcement Learning in Safety-Critical Settings. ICML 2020: 11055-11065 - [c191]Suraj Nair, Mohammad Babaeizadeh, Chelsea Finn, Sergey Levine, Vikash Kumar:
TRASS: Time Reversal as Self-Supervision. ICRA 2020: 115-121 - [c190]Akhil Padmanabha, Frederik Ebert, Stephen Tian, Roberto Calandra, Chelsea Finn, Sergey Levine:
OmniTact: A Multi-Directional High-Resolution Touch Sensor. ICRA 2020: 618-624 - [c189]Avi Singh, Eric Jang, Alexander Irpan, Daniel Kappler, Murtaza Dalal, Sergey Levine, Mohi Khansari, Chelsea Finn:
Scalable Multi-Task Imitation Learning with Autonomous Improvement. ICRA 2020: 2167-2173 - [c188]Gokul Swamy, Siddharth Reddy, Sergey Levine, Anca D. Dragan:
Scaled Autonomy: Enabling Human Operators to Control Robot Fleets. ICRA 2020: 5942-5948 - [c187]Gerrit Schoettler, Ashvin Nair, Jianlan Luo, Shikhar Bahl, Juan Aparicio Ojea, Eugen Solowjow, Sergey Levine:
Deep Reinforcement Learning for Industrial Insertion Tasks with Visual Inputs and Natural Rewards. IROS 2020: 5548-5555 - [c186]Gerrit Schoettler, Ashvin Nair, Juan Aparicio Ojea, Sergey Levine, Eugen Solowjow:
Meta-Reinforcement Learning for Robotic Industrial Insertion Tasks. IROS 2020: 9728-9735 - [c185]Michael Dennis, Natasha Jaques, Eugene Vinitsky
, Alexandre M. Bayen, Stuart Russell, Andrew Critch, Sergey Levine:
Emergent Complexity and Zero-shot Transfer via Unsupervised Environment Design. NeurIPS 2020 - [c184]Ben Eysenbach, Xinyang Geng, Sergey Levine, Ruslan Salakhutdinov:
Rewriting History with Inverse RL: Hindsight Inference for Policy Improvement. NeurIPS 2020 - [c183]Michael Janner, Igor Mordatch, Sergey Levine:
Gamma-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction. NeurIPS 2020 - [c182]Aviral Kumar, Abhishek Gupta, Sergey Levine:
DisCor: Corrective Feedback in Reinforcement Learning via Distribution Correction. NeurIPS 2020 - [c181]Saurabh Kumar, Aviral Kumar, Sergey Levine, Chelsea Finn:
One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL. NeurIPS 2020 - [c180]Aviral Kumar, Sergey Levine:
Model Inversion Networks for Model-Based Optimization. NeurIPS 2020 - [c179]Aviral Kumar, Aurick Zhou, George Tucker, Sergey Levine:
Conservative Q-Learning for Offline Reinforcement Learning. NeurIPS 2020 - [c178]Alex X. Lee, Anusha Nagabandi, Pieter Abbeel, Sergey Levine:
Stochastic Latent Actor-Critic: Deep Reinforcement Learning with a Latent Variable Model. NeurIPS 2020 - [c177]Karl Pertsch, Oleh Rybkin, Frederik Ebert, Shenghao Zhou, Dinesh Jayaraman, Chelsea Finn, Sergey Levine:
Long-Horizon Visual Planning with Goal-Conditioned Hierarchical Predictors. NeurIPS 2020 - [c176]Kelvin Xu, Siddharth Verma, Chelsea Finn, Sergey Levine:
Continual Learning of Control Primitives : Skill Discovery via Reset-Games. NeurIPS 2020 - [c175]Tianhe Yu, Saurabh Kumar, Abhishek Gupta, Sergey Levine, Karol Hausman, Chelsea Finn:
Gradient Surgery for Multi-Task Learning. NeurIPS 2020 - [c174]Tianhe Yu, Garrett Thomas, Lantao Yu, Stefano Ermon, James Y. Zou, Sergey Levine, Chelsea Finn, Tengyu Ma:
MOPO: Model-based Offline Policy Optimization. NeurIPS 2020 - [c173]Xue Bin Peng, Erwin Coumans, Tingnan Zhang, Tsang-Wei Edward Lee, Jie Tan, Sergey Levine:
Learning Agile Robotic Locomotion Skills by Imitating Animals. Robotics: Science and Systems 2020 - [c172]Archit Sharma, Michael Ahn, Sergey Levine, Vikash Kumar, Karol Hausman, Shixiang Gu:
Emergent Real-World Robotic Skills via Unsupervised Off-Policy Reinforcement Learning. Robotics: Science and Systems 2020 - [c171]Laura M. Smith, Nikita Dhawan, Marvin Zhang, Pieter Abbeel, Sergey Levine:
AVID: Learning Multi-Stage Tasks via Pixel-Level Translation of Human Videos. Robotics: Science and Systems 2020 - [i258]Tianhe Yu, Saurabh Kumar, Abhishek Gupta, Sergey Levine, Karol Hausman, Chelsea Finn:
Gradient Surgery for Multi-Task Learning. CoRR abs/2001.06782 (2020) - [i257]Gregory Kahn, Pieter Abbeel, Sergey Levine:
BADGR: An Autonomous Self-Supervised Learning-Based Navigation System. CoRR abs/2002.05700 (2020) - [i256]Sehoon Ha, Peng Xu, Zhenyu Tan, Sergey Levine, Jie Tan:
Learning to Walk in the Real World with Minimal Human Effort. CoRR abs/2002.08550 (2020) - [i255]Benjamin Eysenbach, Xinyang Geng, Sergey Levine, Ruslan Salakhutdinov:
Rewriting History with Inverse RL: Hindsight Inference for Policy Improvement. CoRR abs/2002.11089 (2020) - [i254]Avi Singh, Eric Jang, Alexander Irpan, Daniel Kappler, Murtaza Dalal, Sergey Levine, Mohi Khansari, Chelsea Finn:
Scalable Multi-Task Imitation Learning with Autonomous Improvement. CoRR abs/2003.02636 (2020) - [i253]Akhil Padmanabha, Frederik Ebert, Stephen Tian, Roberto Calandra, Chelsea Finn, Sergey Levine:
OmniTact: A Multi-Directional High Resolution Touch Sensor. CoRR abs/2003.06965 (2020) - [i252]Aviral Kumar, Abhishek Gupta, Sergey Levine:
DisCor: Corrective Feedback in Reinforcement Learning via Distribution Correction. CoRR abs/2003.07305 (2020) - [i251]Xinshuo Weng, Jianren Wang, Sergey Levine, Kris Kitani, Nicholas Rhinehart:
Unsupervised Sequence Forecasting of 100,000 Points for Unsupervised Trajectory Forecasting. CoRR abs/2003.08376 (2020) - [i250]Xue Bin Peng, Erwin Coumans, Tingnan Zhang, Tsang-Wei Edward Lee, Jie Tan, Sergey Levine:
Learning Agile Robotic Locomotion Skills by Imitating Animals. CoRR abs/2004.00784 (2020) - [i249]Ted Xiao, Eric Jang, Dmitry Kalashnikov, Sergey Levine, Julian Ibarz, Karol Hausman, Alexander Herzog:
Thinking While Moving: Deep Reinforcement Learning with Concurrent Control. CoRR abs/2004.06089 (2020) - [i248]Justin Fu, Aviral Kumar, Ofir Nachum, George Tucker, Sergey Levine:
D4RL: Datasets for Deep Data-Driven Reinforcement Learning. CoRR abs/2004.07219 (2020) - [i247]Ryan Julian, Benjamin Swanson, Gaurav S. Sukhatme, Sergey Levine, Chelsea Finn, Karol Hausman:
Efficient Adaptation for End-to-End Vision-Based Robotic Manipulation. CoRR abs/2004.10190 (2020) - [i246]Suneel Belkhale, Rachel Li, Gregory Kahn, Rowan McAllister, Roberto Calandra, Sergey Levine:
Model-Based Meta-Reinforcement Learning for Flight with Suspended Payloads. CoRR abs/2004.11345 (2020) - [i245]Anirudh Goyal, Yoshua Bengio, Matthew M. Botvinick, Sergey Levine:
The Variational Bandwidth Bottleneck: Stochastic Evaluation on an Information Budget. CoRR abs/2004.11935 (2020) - [i244]Henry Zhu, Justin Yu, Abhishek Gupta, Dhruv Shah, Kristian Hartikainen, Avi Singh, Vikash Kumar, Sergey Levine:
The Ingredients of Real-World Robotic Reinforcement Learning. CoRR abs/2004.12570 (2020) - [i243]Archit Sharma, Michael Ahn, Sergey Levine, Vikash Kumar, Karol Hausman, Shixiang Gu:
Emergent Real-World Robotic Skills via Unsupervised Off-Policy Reinforcement Learning. CoRR abs/2004.12974 (2020) - [i242]Gerrit Schoettler, Ashvin Nair, Juan Aparicio Ojea, Sergey Levine, Eugen Solowjow:
Meta-Reinforcement Learning for Robotic Industrial Insertion Tasks. CoRR abs/2004.14404 (2020) - [i241]Sergey Levine, Aviral Kumar, George Tucker, Justin Fu:
Offline Reinforcement Learning: Tutorial, Review, and Perspectives on Open Problems. CoRR abs/2005.01643 (2020) - [i240]Tianhe Yu, Garrett Thomas, Lantao Yu, Stefano Ermon, James Zou, Sergey Levine, Chelsea Finn, Tengyu Ma:
MOPO: Model-based Offline Policy Optimization. CoRR abs/2005.13239 (2020) - [i239]Aviral Kumar, Aurick Zhou, George Tucker, Sergey Levine:
Conservative Q-Learning for Offline Reinforcement Learning. CoRR abs/2006.04779 (2020) - [i238]Russell Mendonca, Xinyang Geng, Chelsea Finn, Sergey Levine:
Meta-Reinforcement Learning Robust to Distributional Shift via Model Identification and Experience Relabeling. CoRR abs/2006.07178 (2020) - [i237]Kanishka Rao, Chris Harris, Alex Irpan, Sergey Levine, Julian Ibarz, Mohi Khansari:
RL-CycleGAN: Reinforcement Learning Aware Simulation-To-Real. CoRR abs/2006.09001 (2020) - [i236]Ashvin Nair, Murtaza Dalal, Abhishek Gupta, Sergey Levine:
Accelerating Online Reinforcement Learning with Offline Datasets. CoRR abs/2006.09359 (2020) - [i235]Amy Zhang
, Rowan McAllister, Roberto Calandra, Yarin Gal, Sergey Levine:
Learning Invariant Representations for Reinforcement Learning without Reconstruction. CoRR abs/2006.10742 (2020) - [i234]John D. Co-Reyes, Suvansh Sanjeev, Glen Berseth, Abhishek Gupta, Sergey Levine:
Ecological Reinforcement Learning. CoRR abs/2006.12478 (2020) - [i233]Oleh Rybkin, Kostas Daniilidis, Sergey Levine:
Simple and Effective VAE Training with Calibrated Decoders. CoRR abs/2006.13202 (2020) - [i232]Karl Pertsch, Oleh Rybkin, Frederik Ebert, Chelsea Finn, Dinesh Jayaraman, Sergey Levine:
Long-Horizon Visual Planning with Goal-Conditioned Hierarchical Predictors. CoRR abs/2006.13205 (2020) - [i231]Benjamin Eysenbach, Swapnil Asawa, Shreyas Chaudhari, Ruslan Salakhutdinov, Sergey Levine:
Off-Dynamics Reinforcement Learning: Training for Transfer with Domain Classifiers. CoRR abs/2006.13916 (2020) - [i230]Angelos Filos, Panagiotis Tigas, Rowan McAllister, Nicholas Rhinehart, Sergey Levine, Yarin Gal:
Can Autonomous Vehicles Identify, Recover From, and Adapt to Distribution Shifts? CoRR abs/2006.14911 (2020) - [i229]Anirudh Goyal, Alex Lamb, Phanideep Gampa, Philippe Beaudoin, Sergey Levine, Charles Blundell, Yoshua Bengio, Michael Mozer:
Object Files and Schemata: Factorizing Declarative and Procedural Knowledge in Dynamical Systems. CoRR abs/2006.16225 (2020) - [i228]Michael Chang, Sidhant Kaushik, S. Matthew Weinberg, Thomas L. Griffiths, Sergey Levine:
Decentralized Reinforcement Learning: Global Decision-Making via Local Economic Transactions. CoRR abs/2007.02382 (2020) - [i227]Marvin Zhang, Henrik Marklund, Abhishek Gupta, Sergey Levine, Chelsea Finn:
Adaptive Risk Minimization: A Meta-Learning Approach for Tackling Group Shift. CoRR abs/2007.02931 (2020) - [i226]Siddharth Reddy, Sergey Levine, Anca D. Dragan:
Assisted Perception: Optimizing Observations to Communicate State. CoRR abs/2008.02840 (2020) - [i225]Eric Mitchell, Rafael Rafailov, Xue Bin Peng, Sergey Levine, Chelsea Finn:
Offline Meta-Reinforcement Learning with Advantage Weighting. CoRR abs/2008.06043 (2020) - [i224]Jesse Zhang, Brian Cheung, Chelsea Finn, Sergey Levine, Dinesh Jayaraman:
Cautious Adaptation For Reinforcement Learning in Safety-Critical Settings. CoRR abs/2008.06622 (2020) - [i223]Kamal Ndousse, Douglas Eck, Sergey Levine, Natasha Jaques
:
Multi-agent Social Reinforcement Learning Improves Generalization. CoRR abs/2010.00581 (2020) - [i222]Gregory Kahn, Pieter Abbeel, Sergey Levine:
LaND: Learning to Navigate from Disengagements. CoRR abs/2010.04689 (2020) - [i221]Anurag Ajay, Aviral Kumar, Pulkit Agrawal, Sergey Levine, Ofir Nachum:
OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning. CoRR abs/2010.13611 (2020) - [i220]Tony Z. Zhao, Anusha Nagabandi, Kate Rakelly, Chelsea Finn, Sergey Levine:
MELD: Meta-Reinforcement Learning from Images via Latent State Models. CoRR abs/2010.13957 (2020) - [i219]Saurabh Kumar, Aviral Kumar, Sergey Levine, Chelsea Finn:
One Solution is Not All You Need: Few-Shot Extrapolation via Structured MaxEnt RL. CoRR abs/2010.14484 (2020) - [i218]Michael Janner, Igor Mordatch, Sergey Levine:
γ-Models: Generative Temporal Difference Learning for Infinite-Horizon Prediction. CoRR abs/2010.14496 (2020) - [i217]Homanga Bharadhwaj, Aviral Kumar, Nicholas Rhinehart, Sergey Levine, Florian Shkurti, Animesh Garg:
Conservative Safety Critics for Exploration. CoRR abs/2010.14497 (2020) - [i216]Aviral Kumar, Rishabh Agarwal, Dibya Ghosh, Sergey Levine:
Implicit Under-Parameterization Inhibits Data-Efficient Deep Reinforcement Learning. CoRR abs/2010.14498 (2020) - [i215]Avi Singh, Albert Yu, Jonathan Yang, Jesse Zhang, Aviral Kumar, Sergey Levine:
COG: Connecting New Skills to Past Experience with Offline Reinforcement Learning. CoRR abs/2010.14500 (2020) - [i214]Dhruv Batra, Angel X. Chang, Sonia Chernova, Andrew J. Davison, Jia Deng, Vladlen Koltun, Sergey Levine, Jitendra Malik, Igor Mordatch, Roozbeh Mottaghi, Manolis Savva, Hao Su:
Rearrangement: A Challenge for Embodied AI. CoRR abs/2011.01975 (2020) - [i213]Aurick Zhou, Sergey Levine:
Amortized Conditional Normalized Maximum Likelihood. CoRR abs/2011.02696 (2020) - [i212]Kelvin Xu, Siddharth Verma, Chelsea Finn, Sergey Levine:
Continual Learning of Control Primitives: Skill Discovery via Reset-Games. CoRR abs/2011.05286 (2020) - [i211]Karl Schmeckpeper, Oleh Rybkin, Kostas Daniilidis, Sergey Levine, Chelsea Finn:
Reinforcement Learning with Videos: Combining Offline Observations with Interaction. CoRR abs/2011.06507 (2020) - [i210]Benjamin Eysenbach, Ruslan Salakhutdinov, Sergey Levine:
C-Learning: Learning to Achieve Goals via Recursive Classification. CoRR abs/2011.08909 (2020) - [i209]Avi Singh, Huihan Liu, Gaoyue Zhou, Albert Yu, Nicholas Rhinehart, Sergey Levine:
Parrot: Data-Driven Behavioral Priors for Reinforcement Learning. CoRR abs/2011.10024 (2020) - [i208]Michael Dennis, Natasha Jaques
, Eugene Vinitsky, Alexandre M. Bayen, Stuart Russell, Andrew Critch, Sergey Levine:
Emergent Complexity and Zero-shot Transfer via Unsupervised Environment Design. CoRR abs/2012.02096 (2020) - [i207]Mohammad Babaeizadeh, Mohammad Taghi Saffar, Danijar Hafner, Harini Kannan, Chelsea Finn, Sergey Levine, Dumitru Erhan:
Models, Pixels, and Rewards: Evaluating Design Trade-offs in Visual Model-Based Reinforcement Learning. CoRR abs/2012.04603 (2020) - [i206]Pang Wei Koh, Shiori Sagawa, Henrik Marklund, Sang Michael Xie, Marvin Zhang, Akshay Balsubramani, Weihua Hu, Michihiro Yasunaga, Richard Lanas Phillips, Sara Beery, Jure Leskovec, Anshul Kundaje, Emma Pierson, Sergey Levine, Chelsea Finn, Percy Liang:
WILDS: A Benchmark of in-the-Wild Distribution Shifts. CoRR abs/2012.07421 (2020) - [i205]Tianhe Yu, Xinyang Geng, Chelsea Finn, Sergey Levine:
Variable-Shot Adaptation for Online Meta-Learning. CoRR abs/2012.07769 (2020) - [i204]Dhruv Shah, Benjamin Eysenbach, Gregory Kahn, Nicholas Rhinehart, Sergey Levine:
ViNG: Learning Open-World Navigation with Visual Goals. CoRR abs/2012.09812 (2020) - [i203]Stephen Tian, Suraj Nair, Frederik Ebert, Sudeep Dasari, Benjamin Eysenbach, Chelsea Finn, Sergey Levine:
Model-Based Visual Planning with Self-Supervised Functional Distances. CoRR abs/2012.15373 (2020)
2010 – 2019
- 2019
- [j11]Nathan O. Lambert
, Daniel S. Drew
, Joseph Yaconelli, Sergey Levine, Roberto Calandra
, Kristofer S. J. Pister
:
Low-Level Control of a Quadrotor With Deep Model-Based Reinforcement Learning. IEEE Robotics Autom. Lett. 4(4): 4224-4230 (2019) - [c170]Ashvin Nair, Shikhar Bahl, Alexander Khazatsky, Vitchyr Pong, Glen Berseth, Sergey Levine:
Contextual Imagined Goals for Self-Supervised Robotic Learning. CoRL 2019: 530-539 - [c169]Sudeep Dasari, Frederik Ebert, Stephen Tian, Suraj Nair, Bernadette Bucher, Karl Schmeckpeper, Siddharth Singh, Sergey Levine, Chelsea Finn:
RoboNet: Large-Scale Multi-Robot Learning. CoRL 2019: 885-897 - [c168]Abhishek Gupta, Vikash Kumar, Corey Lynch, Sergey Levine, Karol Hausman:
Relay Policy Learning: Solving Long-Horizon Tasks via Imitation and Reinforcement Learning. CoRL 2019: 1025-1037 - [c167]Tianhe Yu, Deirdre Quillen, Zhanpeng He, Ryan Julian, Karol Hausman, Chelsea Finn, Sergey Levine:
Meta-World: A Benchmark and Evaluation for Multi-Task and Meta Reinforcement Learning. CoRL 2019: 1094-1100 - [c166]Anusha Nagabandi, Kurt Konolige, Sergey Levine, Vikash Kumar:
Deep Dynamics Models for Learning Dexterous Manipulation. CoRL 2019: 1101-1112 - [c165]Corey Lynch, Mohi Khansari, Ted Xiao, Vikash Kumar, Jonathan Tompson, Sergey Levine, Pierre Sermanet:
Learning Latent Plans from Play. CoRL 2019: 1113-1132 - [c164]Michael Ahn, Henry Zhu, Kristian Hartikainen, Hugo Ponte, Abhishek Gupta, Sergey Levine, Vikash Kumar:
ROBEL: Robotics Benchmarks for Learning with Low-Cost Robots. CoRL 2019: 1300-1313 - [c163]Rishi Veerapaneni, John D. Co-Reyes, Michael Chang, Michael Janner, Chelsea Finn, Jiajun Wu, Joshua B. Tenenbaum, Sergey Levine:
Entity Abstraction in Visual Model-Based Reinforcement Learning. CoRL 2019: 1439-1456 - [c162]Stephen James, Paul Wohlhart, Mrinal Kalakrishnan, Dmitry Kalashnikov, Alex Irpan, Julian Ibarz, Sergey Levine, Raia Hadsell, Konstantinos Bousmalis:
Sim-To-Real via Sim-To-Sim: Data-Efficient Robotic Grasping via Randomized-To-Canonical Adaptation Networks. CVPR 2019: 12627-12637 - [c161]Nicholas Rhinehart
, Rowan McAllister, Kris Kitani, Sergey Levine:
PRECOG: PREdiction Conditioned on Goals in Visual Multi-Agent Settings. ICCV 2019: 2821-2830 - [c160]Michael Chang, Abhishek Gupta, Sergey Levine, Thomas L. Griffiths:
Automatically Composing Representation Transformations as a Means for Generalization. ICLR (Poster) 2019 - [c159]John D. Co-Reyes, Abhishek Gupta, Suvansh Sanjeev, Nick Altieri, Jacob Andreas, John DeNero, Pieter Abbeel, Sergey Levine:
Guiding Policies with Language via Meta-Learning. ICLR (Poster) 2019 - [c158]Benjamin Eysenbach, Abhishek Gupta, Julian Ibarz, Sergey Levine:
Diversity is All You Need: Learning Skills without a Reward Function. ICLR (Poster) 2019 - [c157]Justin Fu, Anoop Korattikara, Sergey Levine, Sergio Guadarrama:
From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following. ICLR (Poster) 2019 - [c156]Dibya Ghosh, Abhishek Gupta, Sergey Levine:
Learning Actionable Representations with Goal Conditioned Policies. ICLR (Poster) 2019 - [c155]Anirudh Goyal, Philemon Brakel, William Fedus, Soumye Singhal, Timothy P. Lillicrap, Sergey Levine, Hugo Larochelle, Yoshua Bengio:
Recall Traces: Backtracking Models for Efficient Reinforcement Learning. ICLR (Poster) 2019 - [c154]Anirudh Goyal, Riashat Islam, Daniel Strouse, Zafarali Ahmed, Hugo Larochelle, Matthew M. Botvinick, Yoshua Bengio, Sergey Levine:
InfoBot: Transfer and Exploration via the Information Bottleneck. ICLR (Poster) 2019 - [c153]