ResearchI specialize in reinforcement learning with a focus on imitation learning (IL). My research primarily concerns improving the sample efficiency of IL. I utilize ideas from statistical learning theory and optimization theory to analyze and design IL methods. Background:Imitation learning (IL) trains good policies from expert demonstrations, and it has been applied in various domains such as robotics and recommendation systems. Contribution:
Reference:[1] Xu, T., Li, Z., and Yu, Y. Error Bounds of Imitating Policies and Environments. NeurIPS 2020. [2] Xu, T., Li, Z., and Yu, Y. Error Bounds of Imitating Policies and Environments for Reinforcement Learning. TPAMI 2021. [3] Li, Z., Xu, T., Yu, Y., and Luo, Z.-Q. Rethinking ValueDice: Does It Really Improve Performance? ICLR 2021. [4] Xu, T., Li, Z., Yu, Y., and Luo, Z.-Q. Understanding Adversarial Imitation Learning in Small Sample Regime: A Stage-coupled Analysis. arXiv:2208.01899. [5] Xu, T., Li, Z., and Yu, Y. Provably Efficient Adversarial Imitation Learning with Unknown Transitions. UAI 2023. [6] Li, Z., Xu, T., Yu, Y., and Luo, Z.-Q. Imitation Learning from Imperfection: Theoretical Justifications and Algorithms. NeurIPS 2023. [7] Xu, T., Zhang, Z., Chen, R., Sun, Y., and Yu, Y. Provably and Practically Efficient Adversarial Imitation Learning with General Function Approximation. NeurIPS 2024 |