Seguir
Mengdi Xu
Mengdi Xu
Dirección de correo verificada de stanford.edu - Página principal
Título
Citado por
Citado por
Año
Prompting decision transformer for few-shot policy generalization
M Xu, Y Shen, S Zhang, Y Lu, D Zhao, J Tenenbaum, C Gan
international conference on machine learning, 24631-24645, 2022
1322022
Delay-aware model-based reinforcement learning for continuous control
B Chen, M Xu, L Li, D Zhao
Neurocomputing 450, 119-128, 2021
742021
Cmts: A conditional multiple trajectory synthesizer for generating safety-critical driving scenarios
W Ding, M Xu, D Zhao
2020 IEEE International Conference on Robotics and Automation (ICRA), 4314-4321, 2020
622020
Trustworthy reinforcement learning against intrinsic vulnerabilities: Robustness, safety, and generalizability
M Xu, Z Liu, P Huang, W Ding, Z Cen, B Li, D Zhao
arXiv preprint arXiv:2209.08025, 2022
472022
Hyper-decision transformer for efficient online policy adaptation
M Xu, Y Lu, Y Shen, S Zhang, D Zhao, C Gan
arXiv preprint arXiv:2304.08487, 2023
422023
Context-Aware Safe Reinforcement Learning for Non-Stationary Environments
B Chen, Z Liu, J Zhu, M Xu, W Ding, D Zhao
2021 IEEE International Conference on Robotics and Automation (ICRA), 2021
392021
Task-agnostic online reinforcement learning with an infinite mixture of gaussian processes
M Xu, W Ding, J Zhu, Z Liu, B Chen, D Zhao
Advances in Neural Information Processing Systems 33, 6429-6440, 2020
382020
Creative robot tool use with large language models
M Xu, P Huang, W Yu, S Liu, X Zhang, Y Niu, T Zhang, F Xia, J Tan, ...
arXiv preprint arXiv:2310.13065, 2023
362023
Robust Reinforcement Learning as a Stackelberg Game via Adaptively-Regularized Adversarial Training
P Huang, M Xu, F Fang, D Zhao
31st International Joint Conference on Artificial Intelligence (IJCAI 2022), 2022
362022
A review of energy supply for biomachine hybrid robots
Z Ma, J Zhao, L Yu, M Yan, L Liang, X Wu, M Xu, W Wang, S Yan
Cyborg and Bionic Systems 4, 0053, 2023
342023
Delay-aware multi-agent reinforcement learning for cooperative and competitive environments
B Chen, M Xu, Z Liu, L Li, D Zhao
arXiv preprint arXiv:2005.05441, 2020
342020
Curriculum Reinforcement Learning using Optimal Transport via Gradual Domain Adaptation
P Huang, M Xu, J Zhu, L Shi, F Fang, D Zhao
NeurIPS 2022, 2022
292022
Can brain signals reveal inner alignment with human languages?
J Qiu, W Han, J Zhu, M Xu, D Weber, B Li, D Zhao
Findings of the Association for Computational Linguistics: EMNLP 2023, 1789-1804, 2023
18*2023
Cardiac Disease Diagnosis On Imbalanced Electrocardiography Data Through Optimal Transport Augmentation
J Qiu, J Zhu, M Xu, P Huang, M Rosenberg, D Weber, E Liu, D Zhao
2023 IEEE International Conference on Acoustics, Speech, and Signal Processing, 2023
18*2023
What went wrong? closing the sim-to-real gap via differentiable causal discovery
P Huang, X Zhang, Z Cao, S Liu, M Xu, W Ding, J Francis, B Chen, ...
Conference on Robot Learning, 734-760, 2023
172023
Transfer knowledge from natural language to electrocardiography: Can we detect cardiovascular disease through language models?
J Qiu, W Han, J Zhu, M Xu, M Rosenberg, E Liu, D Weber, D Zhao
arXiv preprint arXiv:2301.09017, 2023
172023
Mhms: Multimodal hierarchical multimedia summarization
J Qiu, J Zhu, M Xu, F Dernoncourt, T Bui, Z Wang, B Li, D Zhao, H Jin
arXiv preprint arXiv:2204.03734, 2022
172022
Scalable Safety-Critical Policy Evaluation with Accelerated Rare Event Sampling
M Xu, P Huang, F Li, J Zhu, X Qi, K Oguchi, Z Huang, H Lam, D Zhao
The 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems …, 2021
16*2021
Adaptive online replanning with diffusion models
S Zhou, Y Du, S Zhang, M Xu, Y Shen, W Xiao, DY Yeung, C Gan
Advances in Neural Information Processing Systems 36, 2024
142024
Group distributionally robust reinforcement learning with hierarchical latent variables
M Xu, P Huang, Y Niu, V Kumar, J Qiu, C Fang, KH Lee, X Qi, H Lam, B Li, ...
International Conference on Artificial Intelligence and Statistics, 2677-2703, 2023
112023
El sistema no puede realizar la operación en estos momentos. Inténtalo de nuevo más tarde.
Artículos 1–20