Seguir
Chuheng Zhang
Chuheng Zhang
Microsoft Research
Dirección de correo verificada de microsoft.com
Título
Citado por
Citado por
Año
Exploration by maximizing Rényi entropy for reward-free RL framework
C Zhang, Y Cai, L Huang, J Li
Proceedings of the AAAI Conference on Artificial Intelligence 35 (12), 10859 …, 2021
492021
Return-based contrastive representation learning for reinforcement learning
G Liu, C Zhang, L Zhao, T Qin, J Zhu, J Li, N Yu, TY Liu
arXiv preprint arXiv:2102.10960, 2021
492021
Observation of topological links associated with Hopf insulators in a solid-state quantum simulator
XX Yuan, L He, ST Wang, DL Deng, F Wang, WQ Lian, X Wang, ...
Chinese Physics Letters 34 (6), 060302, 2017
392017
Cross DQN: Cross deep Q network for ads allocation in feed
G Liao, Z Wang, X Wu, X Shi, C Zhang, Y Wang, X Wang, D Wang
Proceedings of the ACM Web Conference 2022, 401-409, 2022
322022
Inductive matrix completion using graph autoencoder
W Shen, C Zhang, Y Tian, L Zeng, X He, W Dou, X Xu
Proceedings of the 30th ACM International Conference on Information …, 2021
222021
Auxiliary-task based deep reinforcement learning for participant selection problem in mobile crowdsourcing
W Shen, X He, C Zhang, Q Ni, W Dou, Y Wang
Proceedings of the 29th ACM International Conference on Information …, 2020
192020
DoubleEnsemble: A new ensemble method based on sample reweighting and feature selection for financial data analysis
C Zhang, Y Li, X Chen, Y Jin, P Tang, J Li
2020 IEEE International Conference on Data Mining (ICDM), 781-790, 2020
182020
Multi-agent reinforcement learning with shared resources for inventory management
Y Ding, M Feng, G Liu, W Jiang, C Zhang, L Zhao, L Song, H Li, Y Jin, ...
arXiv preprint arXiv:2212.07684, 2022
162022
Pre-trained large language models for industrial control
L Song, C Zhang, L Zhao, J Bian
arXiv preprint arXiv:2308.03028, 2023
82023
A versatile multi-agent reinforcement learning benchmark for inventory management
X Yang, Z Liu, W Jiang, C Zhang, L Zhao, L Song, J Bian
arXiv preprint arXiv:2306.07542, 2023
82023
Deep page-level interest network in reinforcement learning for ads allocation
G Liao, X Shi, Z Wang, X Wu, C Zhang, Y Wang, X Wang, D Wang
Proceedings of the 45th International ACM SIGIR Conference on Research and …, 2022
82022
Policy Search by Target Distribution Learning for Continuous Control.
C Zhang, Y Li, J Li
AAAI, 6770-6777, 2020
82020
Venlafaxine as an adjuvant therapy for inflammatory bowel disease patients with anxious and depressive symptoms: a randomized controlled trial
C Liang, P Chen, Y Tang, C Zhang, N Lei, Y Luo, S Duan, Y Zhang
Frontiers in Psychiatry 13, 880058, 2022
72022
Towards generalizable reinforcement learning for trade execution
C Zhang, Y Duan, X Chen, J Chen, J Li, L Zhao
arXiv preprint arXiv:2307.11685, 2023
52023
Whittle Index with Multiple Actions and State Constraint for Inventory Management
C Zhang, X Wang, W Jiang, X Yang, S Wang, L Song, J Bian
The Twelfth International Conference on Learning Representations, 0
5
RePreM: representation pre-training with masked model for reinforcement learning
Y Cai, C Zhang, W Shen, X Zhang, W Ruan, L Huang
Proceedings of the AAAI Conference on Artificial Intelligence 37 (6), 6879-6887, 2023
42023
Learning List-wise Representation in Reinforcement Learning for Ads Allocation with Multiple Auxiliary Tasks
Z Wang, G Liao, X Shi, X Wu, C Zhang, Y Wang, X Wang, D Wang
Proceedings of the 31st ACM International Conference on Information …, 2022
42022
Empowering Large Language Models on Robotic Manipulation with Affordance Prompting
G Cheng, C Zhang, W Cai, L Zhao, C Sun, J Bian
arXiv preprint arXiv:2404.11027, 2024
32024
Imitation learning to outperform demonstrators by directly extrapolating demonstrations
Y Cai, C Zhang, W Shen, X He, X Zhang, L Huang
Proceedings of the 31st ACM International Conference on Information …, 2022
32022
Massively parallel market simulator for financial reinforcement learning
J Han, Z Xia, XY Liu, C Zhang, Z Wang, J Guo
AI in Finance Bridge, AAAI, 2023
22023
El sistema no puede realizar la operación en estos momentos. Inténtalo de nuevo más tarde.
Artículos 1–20