[1] LEVINE S, KUMAR A, TUCKER G, et al. Offline reinforcement learning: tutorial, review, and perspectives on open problems[J]. arXiv:2005.01643, 2020.
[2] OKUMURA B, JAMES M R, KANZAWA Y, et al. Challenges in perception and decision making for intelligent automotive vehicles: a case study[J]. IEEE Transactions on Intelligent Vehicles, 2016, 1(1): 20-32.
[3] WANG P, GAO S, LI L, et al. Obstacle avoidance path planning design for autonomous driving vehicles based on an improved artificial potential field algorithm[J]. Energies, 2019, 12(12): 2342.
[4] CUI Q, DING R, WEI C, et al. A hierarchical framework of emergency collision avoidance amid surrounding vehicles in highway driving[J]. Control Engineering Practice, 2021, 109: 104751.
[5] ZHANG Y, ZHANG J, ZHANG J, et al. A novel learning framework for sampling-based motion planning in autonomous driving[C]//Proceedings of the AAAI Conference on Artificial Intelligence, 2020: 1202-1209.
[6] AHN J, KIM M, PARK J. Vision-based autonomous driving for unstructured environments using imitation learning[J]. arXiv:2202.10002, 2022.
[7] ZHANG L, ZHANG R, WU T, et al. Safe reinforcement learning with stability guarantee for motion planning of autonomous vehicles[J]. IEEE Transactions on Neural Networks and Learning Systems, 2021, 32(12): 5435-5444.
[8] SHU K, YU H, CHEN X, et al. Autonomous driving at intersections: a behavior-oriented critical-turning-point approach for decision making[J]. IEEE/ASME Transactions on Mechatronics, 2021, 27(1): 234-244.
[9] HANG P, LV C, HUANG C, et al. Cooperative decision making of connected automated vehicles at multi-lane merging zone: a coalitional game approach[J]. IEEE Transactions on Intelligent Transportation Systems, 2021, 23(4): 3829-3841.
[10] FUJIMOTO S, MEGER D, PRECUP D. Off-policy deep reinforcement learning without exploration[C]//International Conference on Machine Learning, 2019: 2052-2062.
[11] FUJIMOTO S, GU S S. A minimalist approach to offline reinforcement learning[C]//Advances in Neural Information Processing Systems, 2021: 20132-20145.
[12] KUMAR A, ZHOU A, TUCKER G, et al. Conservative q-learning for offline reinforcement learning[C]//Advances in Neural Information Processing Systems, 2020: 1179-1191.
[13] YU T, THOMAS G, YU L, et al. Mopo: model-based offline policy optimization[C]//Advances in Neural Information Processing Systems, 2020: 14129-14142.
[14] VASWANI A, SHAZEER N, PARMAR N, et al. Attention is all you need[C]//Advances in Neural Information Processing Systems, 2017.
[15] LEE J D M C K, TOUTANOVA K. Pre-training of deep bidirectional transformers for language understanding[J]. arXiv:1810.04805, 2018.
[16] LIU Z, LIN Y, CAO Y, et al. Swin transformer: hierarchical vision transformer using shifted windows[C]//Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021: 10012-10022.
[17] DOSOVITSKIY A, BEYER L, KOLESNIKOV A, et al. An image is worth 16x16 words: Transformers for image recognition at scale[J]. arXiv:2010.11929, 2020.
[18] CARION N, MASSA F, SYNNAEVE G, et al. End-to-end object detection with transformers[C]//Proceedings of 16th European Conference on Computer Vision, Glasgow, UK, August 23-28, 2020. [S.l.]: Springer International Publishing, 2020: 213-229.
[19] 张英俊, 白小辉, 谢斌红. CNN-Transformer特征融合多目标跟踪算法[J]. 计算机工程与应用, 2024, 60(2): 180-190.
ZHANG Y J, BAI X H, XIE B H. Multi-object tracking algorithm based on CNN-Transformer feature fusion[J]. Computer Engineering and Applications, 2024, 60(2): 180-190.
[20] 方思严, 刘斌. 小波分频自注意力Transformer图像去雨网络[J]. 计算机工程与应用, 2024, 60(6): 259-273.
FANG S Y, LIU B. Wavelet frequency division self-attention Transformer image deraining network[J]. Computer Engineering and Applications, 2024, 60(6): 259-273.
[21] CHEN L, LU K, RAJESWARAN A, et al. Decision transformer: reinforcement learning via sequence modeling[C]// Advances in Neural Information Processing Systems, 2021: 15084-15097.
[22] RADFORD A, WU J, CHILD R, et al. Language models are unsupervised multitask learners[J]. OpenAI Blog, 2019, 1(8): 9.
[23] BROWN T, MANN B, RYDER N, et al. Language models are few-shot learners[C]//Advances in Neural Information Processing Systems, 2020: 1877-1901.
[24] VAN SEIJEN H, FATEMI M, ROMOFF J, et al. Hybrid reward architecture for reinforcement learning[C]//Advances in Neural Information Processing Systems, 2017.
[25] LI Q, PENG Z, FENG L, et al. Metadrive: composing diverse driving scenarios for generalizable reinforcement learning[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2022: 3461-3475.
[26] XUE Z, PENG Z, LI Q, et al. Guarded policy optimization with imperfect online demonstrations[C]//Proceedings of the Eleventh International Conference on Learning Representations, 2022.
[27] LI Q, PENG Z, WU H, et al. Human-AI shared control via policy dissection[C]//Advances in Neural Information Processing Systems, 2022: 8853-8867.
[28] FU J, KUMAR A, NACHUM O, et al. D4rl: datasets for deep data-driven reinforcement learning[J]. arXiv:2004.07219, 2020.
[29] LIU H, HUANG Z, MO X, et al. Augmenting reinforcement learning with Transformer-based scene representation learning for decision-making of autonomous driving[J]. arXiv:2208.12263, 2022.
[30] FANG X, ZHANG Q, GAO Y, et al. Offline reinforcement learning for autonomous driving with real world driving data[C]//2022 IEEE 25th International Conference on Intelligent Transportation Systems (ITSC), 2022: 3417-3422.
[31] SHI T, CHEN D, CHEN K, et al. Offline reinforcement learning for autonomous driving with safety and exploration enhancement[J]. arXiv:2110.07067, 2021. |