Distributed learning and multi-objectivity in traffic light control

被引:44
作者
Brys, Tim [1 ]
Pham, Tong T. [2 ]
Taylor, Matthew E. [3 ]
机构
[1] Vrije Univ Brussel, Dept Comp Sci, Brussels, Belgium
[2] Lafayette Coll, Dept Comp Sci, Easton, PA 18042 USA
[3] Washington State Univ, Sch Elect Engn & Comp Sci, Pullman, WA 99164 USA
关键词
traffic control; DCEE; reinforcement learning; multi-objective optimisation;
D O I
10.1080/09540091.2014.885282
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Traffic jams and suboptimal traffic flows are ubiquitous in modern societies, and they create enormous economic losses each year. Delays at traffic lights alone account for roughly 10% of all delays in US traffic. As most traffic light scheduling systems currently in use are static, set up by human experts rather than being adaptive, the interest in machine learning approaches to this problem has increased in recent years. Reinforcement learning (RL) approaches are often used in these studies, as they require little pre-existing knowledge about traffic flows. Distributed constraint optimisation approaches (DCOP) have also been shown to be successful, but are limited to cases where the traffic flows are known. The distributed coordination of exploration and exploitation (DCEE) framework was recently proposed to introduce learning in the DCOP framework. In this paper, we present a study of DCEE and RL techniques in a complex simulator, illustrating the particular advantages of each, comparing them against standard isolated traffic actuated signals. We analyse how learning and coordination behave under different traffic conditions, and discuss the multi-objective nature of the problem. Finally we evaluate several alternative reward signals in the best performing approach, some of these taking advantage of the correlation between the problem-inherent objectives to improve performance.
引用
收藏
页码:65 / 83
页数:20
相关论文
共 19 条
[1]  
Albus J. S., 1981, BRAINS BEHAV ROBOTIC
[2]  
[Anonymous], 1994, CITESEER
[3]   A comprehensive survey of multiagent reinforcement learning [J].
Busoniu, Lucian ;
Babuska, Robert ;
De Schutter, Bart .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02) :156-172
[4]  
Chang Y. H., 2004, ADV NEURAL INFORM PR, V16
[5]   A multiagent approach to autonomous intersection management [J].
Dresner, Kurt ;
Stone, Peter .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2008, 31 :591-656
[6]   Multiagent Reinforcement Learning for Integrated Network of Adaptive Traffic Signal Controllers (MARLIN-ATSC): Methodology and Large-Scale Application on Downtown Toronto [J].
El-Tantawy, Samah ;
Abdulhai, Baher ;
Abdelgawad, Hossam .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2013, 14 (03) :1140-1150
[7]  
Klopf AH, 1972, Technical Report No.: AFCRL-72-0164
[8]  
Kuyer L, 2008, LECT NOTES ARTIF INT, V5211, P656, DOI 10.1007/978-3-540-87479-9_61
[9]  
Liu ZY, 2007, INT J COMPUT SCI NET, V7, P105
[10]  
Ma Shou-feng, 2002, Journal of Systems Engineering, V17, P526