Traffic signal control using reinforcement learning with webster method웹스터 방법과 결합한 강화학습을 이용한 교통신호제어

Cited 0 time in webofscience Cited 0 time in scopus
  • Hit : 485
  • Download : 0
Conventional traffic signal control algorithms are static, linear, fuzzy, or non-deterministic. Therefore, they are inappropriate for dealing with transportation systems efficiently. However, the Webster method is one conventional method that can find the optimal solution in one junction. Model predictive control methods have high computing costs in simulation. Machine-learning algorithms have become available as communication technology, the Internet of Things (IoT), and Big Data have improved. One example is reinforcement learning, which has an agent that continually interacts with the environment and is appropriate for data-driven approaches. However, the research on integrating reinforcement learning with traffic signal control usually uses image or matrix data as input, which results in long time requirements for training and difficulty in achieving convergence. Furthermore, a simplified action space is used, such as by expanding the current signal phase or switching to other phases. One model is also trained for one demand scenario, which lacks robustness. Therefore, in this study, the signal control variables of the green ratio and cycle length are defined as actions using a Double Deep-Q Network (DDQN) or the Webster method, and the robustness is improved for various travel patterns. Although reinforcement learning algorithms can find a proper solution in one junction, a trade-off between intersections is required when there are four junctions, which is not possible with the Webster method alone. The results show that with four junctions, the original DDQN and the DDQN initialized with the Webster method provide the best trade-off between intersections and achieve high network performance, but the discrepancy between the two methods is not large. This study contributes to the concept of combining reinforcement learning with traffic signal control variables and algorithms. Further study is needed to develop the concept by applying other reinforcement algorithms or considering other signal control variables such as the offset.
Advisors
Yeo, Hwasooresearcher여화수researcher
Description
한국과학기술원 :건설및환경공학과,
Publisher
한국과학기술원
Issue Date
2018
Identifier
325007
Language
eng
Description

학위논문(석사) - 한국과학기술원 : 건설및환경공학과, 2018.8,[v, 68 p. :]

Keywords

Reinforcement learning▼adouble Deep Q Network▼adata-driven approach▼atraffic signal control▼aITS; 강화학습▼a더블 딥 큐 러닝▼a데이터 기반▼a교통 신호 제어▼a지능형교통시스템

URI
http://hdl.handle.net/10203/265595
Link
http://library.kaist.ac.kr/search/detail/view.do?bibCtrlNo=828406&flag=dissertation
Appears in Collection
CE-Theses_Master(석사논문)
Files in This Item
There are no files associated with this item.

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0