Citation: | HE Feng, XIONG Li, ZHOU Xuan, et al., “Scheduling Pattern of Time Triggered Ethernet Based on Reinforcement Learning,” Chinese Journal of Electronics, vol. 32, no. 5, pp. 1022-1035, 2023, doi: 10.23919/cje.2021.00.419 |
[1] |
IEEE Std 802.3-2022: 1968, IEEE Standard for Ethernet, Available at: https://ieeexplore.ieee.org/document/9844436.
|
[2] |
Y. H. Lee, Safety and Certification Approaches for Ethernet Based Aviation Databuses, Technical Report, DOT/FAA/AR-05/52, Federal Aviation Administration, 2005.
|
[3] |
AEE Committee, Arinc specification 664 p7, ARINC 664 aircraft data network, Avionics full duplex switched Ethernet (AFDX) network, Technical report, Annapolis, MD, USA: Aeronautical Radio Inc., 2005.
|
[4] |
R. I. Davis, A. Burns, R. J. Bril, et al., “Controller area network (CAN) schedulability analysis: refuted, revisited and revised,” Real-Time Systems, vol.35, no.3, pp.239–272, 2005. doi: 10.1007/s11241-007-9012-7
|
[5] |
SAE. SAE AS6802 Time-triggered Ethernet, Warrendale: SAE International, 2011.
|
[6] |
J. D. Decotignie, “Ethernet-based real-time and industrial communications,” Proceedings of the IEEE, vol.93, no.6, pp.1102–1117, 2005. doi: 10.1109/JPROC.2005.849721
|
[7] |
W. Steiner, “An evaluation of SMT-based schedule synthesis for time-triggered multi-hop networks,” in 2010 31st IEEE Real-Time Systems Symposium, San Diego, CA, USA, pp.375–384, 2010.
|
[8] |
V. Mnih, K. Kavukcuoglu, D. Silver, et al., “Human-level control through deep reinforcement learning,” Nature, vol.518, no.7540, pp.529–533, 2015. doi: 10.1038/nature14236
|
[9] |
S. Daftry, J. A. Bagnell, and M. Hebert, “Learning transferable policies for monocular reactive MAV control,” in 15th International Symposium on Experimental Robotics, Nagasaki, Japan, pp.3–11, 2016.
|
[10] |
D. Silver, J. Schrittwieser, K. Simonyan, et al., “Mastering the game of go without human knowledge,” Nature, vol.550, no.7676, pp.354–359, 2017. doi: 10.1038/nature24270
|
[11] |
A. Ecoffet, J. Huizinga, J. Lehman, et al., Go-explore: a new approach for hard-exploration problems, arXiv preprint arXiv: 1901.10995, 2019, doi: 10.48550/arXiv.1901.10995.
|
[12] |
H. J. Dai, E. B. Khalil, Y. Y. Zhang, et al., “Learning combinatorial optimization algorithms over graphs,” in Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach CA, USA, 2017.
|
[13] |
O. Vinyals, M. Fortunato, and N. Jaitly, “Pointer networks,” in Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, Canada, pp.2692–2700, 2015.
|
[14] |
V. Mnih, A. P. Badia, M. Mirza, et al., “Asynchronous methods for deep reinforcement learning,” in Proceedings of the 33rd International Conference on International Conference on Machine Learning, New York, NY, USA, pp.1928–1937, 2016.
|
[15] |
R. Dobrin and G. Fohler, “Implementing off-line message scheduling on controller area network (CAN),” in ETFA 2001. 8th International Conference on Emerging Technologies and Factory Automation, Antibes-Juan les Pins, France, pp.241–245, 2001.
|
[16] |
R. I. Davis and A. Burns, “Robust priority assignment for messages on controller area network (CAN),” Real-Time Systems, vol.41, no.2, pp.152–180, 2009. doi: 10.1007/s11241-008-9065-2
|
[17] |
R. Marau, L. Almeida, P. Pedreiras, et al., “Utilization-based schedulability analysis for switched Ethernet aiming dynamic QoS management,” in 2010 IEEE 15th Conference on Emerging Technologies & Factory Automation, Bilbao, Spain, pp.1–10, 2010.
|
[18] |
E. Suethanuwong, “Scheduling time-triggered traffic in TTEthernet systems,” in Proceedings of 2012 IEEE 17th International Conference on Emerging Technologies & Factory Automation, Krakow, Poland, pp.1–4, 2012.
|
[19] |
F. Glover, “Future paths for integer programming and links to artificial intelligence,” Computers & Operations Research, vol.13, no.5, pp.533–549, 1986. doi: 10.1016/0305-0548(86)90048-1
|
[20] |
Y. J. Zhang, F. He, G. S. Lu, et al., “An imporosity message scheduling based on modified genetic algorithm for time-triggered Ethernet,” Science China Information Sciences, vol.61, article no.019102, 2018. doi: 10.1007/s11432-017-9121-6
|
[21] |
D. Tămaş-Selicean, P. Pop, and W. Steiner, “Design optimization of TTEthernet-based distributed real-time systems,” Real-Time Systems, vol.51, no.1, pp.1–35, 2015. doi: 10.1007/s11241-014-9214-8
|
[22] |
R. E. Korf, “Depth-first iterative-deepening: An optimal admissible tree search,” Artificial Intelligence, vol.27, no.1, pp.97–109, 1985. doi: 10.1016/0004-3702(85)90084-0
|
[23] |
R. S. Sutton, D. A. McAllester, S. P. Singh, et al., “Policy gradient methods for reinforcement learning with function approximation,” in Proceedings of the 12th International Conference on Neural Information Processing Systems, Denver, CO, USA, pp.1057–1063, 1999.
|
[24] |
H. R. Li, F. He, Z. Zheng, et al., “Time-triggered communication scheduling method based on reinforcement learning,” Journal of Beijing University of Aeronautics and Astronautics, vol.45, no.9, pp.1894–1901, 2019. (in Chinese) doi: 10.13700/j.bh.1001-5965.2018.0789
|
[25] |
D. Silver, A. Huang, C. J. Maddison, et al., “Mastering the game of Go with deep neural networks and tree search,” Nature, vol.529, no.7587, pp.484–489, 2016. doi: 10.1038/nature16961
|
[26] |
L. Kocsis and C. Szepesvári, “Bandit based Monte-Carlo planning,” in 17th European Conference on Machine Learning, Berlin, Germany, pp.282–293, 2006.
|
[27] |
D. Silver, G. Lever, N. M. O. Heess, et al., “Deterministic policy gradient algorithms,” in Proceedings of the 31st International Conference on International Conference on Machine Learning, Beijing, China, pp.I-387–I-395, 2014.
|
[28] |
P. Abbeel and A. Y. Ng, “Apprenticeship learning via inverse reinforcement learning,” in Proceedings of the Twenty-First International Conference on Machine Learning, Banff, Canada, pp.1–8, 2004.
|
[29] |
D. Horgan, J. Quan, D. Budden, et al., “Distributed prioritized experience replay,” in 6th International Conference on Learning Representations, Vancouver, Canada, 2018.
|
[30] |
L. X. Zhao, P. Pop, Q. Li, et al., “Timing analysis of rate-constrained traffic in TTEthernet using network calculus,” Real-Time Systems, vol.53, no.2, pp.254–287, 2017. doi: 10.1007/s11241-016-9265-0
|
[31] |
X. Glorot, A. Bordes, and Y. Bengio, “Deep sparse rectifier neural networks,” in Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, pp.315–323, 2011.
|
[32] |
D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in 3rd International Conference on Learning Representations, San Diego, CA, USA, 2015.
|
[33] |
D. Clevert, T. Unterthiner, and S. Hochreiter, “Fast and accurate deep network learning by exponential linear units (ELUs),” in 4th International Conference on Learning Representations, San Juan, Puerto Rico, 2016.
|
[34] |
Y. J. Zhang, F. He, G. S. Lu, et al., “Scheduling rate-constrained flows with dynamic programming priority in time-triggered ethernet,” Chinese Journal of Electronics, vol.26, no.4, pp.849–855, 2017. doi: 10.1049/cje.2017.06.002
|