|
[1]
|
Li, Y., Zhang, W., Wang, C.-X., Sun, J. and Liu, Y. (2020) Deep Reinforcement Learning for Dynamic Spectrum Sens-ing and Aggregation in Multi-Channel Wireless Networks. IEEE Transactions on Cognitive Communications and Net-working, 6, 464-475. [Google Scholar] [CrossRef]
|
|
[2]
|
Poston, J.D. and Horne, W.D. (2005) Discontiguous OFDM Considerations for Dynamic Spectrum Access in Idle TV Channels. First IEEE International Symposium on New Frontiers in Dynamic Spectrum Access Networks, 2005, Baltimore, 8-11 November 2005, 607-610. [Google Scholar] [CrossRef]
|
|
[3]
|
Botsov, M., Klügel, M., Kellerer, W. and Fertl, P. (2014) Location Dependent Resource Allocation for Mobile Device-to-Device Communications. 2014 IEEE Wireless Commu-nications and Networking Conference (WCNC), Istanbul, 6-9 April 2014, 1679-1684. [Google Scholar] [CrossRef]
|
|
[4]
|
Sun, W., Ström, E.G., Brännström, F., Sou, K.C. and Sui, Y. (2016) Radio Resource Management for D2D-Based V2V Communication. IEEE Transactions on Vehicular Technolo-gy, 65, 6636-6650. [Google Scholar] [CrossRef]
|
|
[5]
|
Ye, H., Liang, L., Li, G.Y., Kim, J., Lu, L. and Wu, M. (2018) Machine Learning for Vehicular Networks: Recent Advances and Application Examples. IEEE Vehicular Technology Magazine, 13, 94-101. [Google Scholar] [CrossRef]
|
|
[6]
|
Liang, L., Ye, H. and Li, G.Y. (2019) Toward Intelligent Vehic-ular Networks: A Machine Learning Framework. IEEE Internet of Things Journal, 6, 124-135. [Google Scholar] [CrossRef]
|
|
[7]
|
Liang, L., Ye, H. and Li, G.Y. (2019) Spectrum Sharing in Ve-hicular Networks Based on Multi-Agent Reinforcement Learning. IEEE Journal on Selected Areas in Communications, 37, 2282-2292. [Google Scholar] [CrossRef]
|
|
[8]
|
(2017) Technical Specification Group Radio Access Network. Study Enhancement 3GPP Support for 5G V2X Services. Release 15, Document 3GPP TR 22.886 V15.1.0, 3rd Gener-ation Partnership Project.
|
|
[9]
|
Molina-Masegosa, R. and Gozalvez, J. (2017) LTE-V for Sidelink 5G V2X Vehicular Communications: A New 5G Technology for Short-Range Vehicle-to-Everything Communications. IEEE Vehicular Technology Magazine, 12, 30-39. [Google Scholar] [CrossRef]
|
|
[10]
|
Omidshafiei, S., Pazis, J., Amato, C., How, J.P. and Vian, J. (2017) Deep Decentralized Multi-Task Multi-Agent Reinforcement Learning under Par Tial Observability. Proceedings of the 34th International Conference on Machine Learning, (ICML), Sydney, 6-11 August 2017, 2681-2690.
|
|
[11]
|
Foerster, J., et al. (2017) Stabilising Experience Replay for Deep Multi-Agent Rein-forcement Learning. Proceedings of the 34th International Conference on Machine Learning, Sydney, 6-11 August 2017, 1146-1155.
|
|
[12]
|
Nasir, Y.S. and Guo, D. (2018) Deep Reinforcement Learning for Distributed Dynamic Pow-er Allocation in Wireless Networks. ArXiv: 1808.00490. https://arxiv.org/abs/1808.00490
|
|
[13]
|
Tan, M. (1993) Multi-Agent Reinforcement Learning: Independent vs. Cooperative Agents. Proceedings of the 10th International Con-ference, University of Massachusetts, Amherst, 27-29 June 1993, 330-337. [Google Scholar] [CrossRef]
|
|
[14]
|
Mnih, V., et al. (2015) Human-Level Control through Deep Reinforcement Learning. Nature, 518, 529-533. [Google Scholar] [CrossRef] [PubMed]
|
|
[15]
|
Watkins, C.J.C.H. and Dayan, P. (1992) Q-Learning. Machine Learning, 8, 279-292. [Google Scholar] [CrossRef]
|
|
[16]
|
Sutton, R.S. and Barto, A.G. (1998) Reinforcement Learning: An Intro-duction. MIT Press, Cambridge.
|
|
[17]
|
(2016) Technical Specification Group Radio Access Network. Study LTE-Based V2X Services, Release 14, Document 3GPP TR 36.885 V14.0.0, 3rd Generation Partnership Project.
|
|
[18]
|
Ruder, S. (2016) An Overview of Gradient Descent Optimization Algorithms. ArXiv: 1609.04747.
https://arxiv.org/abs/1609.04747
|
|
[19]
|
Ye, H., Li, G.Y. and Juang, B.-H.F. (2019) Deep Reinforcement Learning Based Resource Allocation for V2V Communications. IEEE Transactions on Vehicular Technology, 68, 3163-3173. [Google Scholar] [CrossRef]
|