Energy-Efficient Online Path Planning of Multiple Drones Using Reinforcement Learning

Cited 32 time in webofscience Cited 0 time in scopus
  • Hit : 250
  • Download : 12
DC FieldValueLanguage
dc.contributor.authorHong, Dooyoungko
dc.contributor.authorLee, Seonhoonko
dc.contributor.authorCho, Young Hooko
dc.contributor.authorBaek, Donkyuko
dc.contributor.authorKim, Jaeminko
dc.contributor.authorChang, Naehyuckko
dc.date.accessioned2021-11-03T06:41:22Z-
dc.date.available2021-11-03T06:41:22Z-
dc.date.created2021-11-02-
dc.date.created2021-11-02-
dc.date.created2021-11-02-
dc.date.issued2021-10-
dc.identifier.citationIEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, v.70, no.10, pp.9725 - 9740-
dc.identifier.issn0018-9545-
dc.identifier.urihttp://hdl.handle.net/10203/288596-
dc.description.abstractDrones, typically unmanned aerial vehicles (UAVs), have many purposes. However, simultaneous operation of multiple drones is challenging, considering the real-time interactions and the environment; the drone must avoid collision with the other drones or obstacles. The proposed Advanced TD3 model performs energy-efficient path planning at the edge-level drone. We modify the twin-delayed deep deterministic policy gradient (TD3), which is the state-of-the-art policy gradient reinforcement learning. The frame stacking technique considers the continuous action space of the drone to the TD3 model. During the training, we gradually increase the observation range of agents for fast and stable convergence. We train the modified TD3 model through Offline RL to reduce the overhead for the RL model training. Drones mount the converged RL model on their onboard computer. The Advanced TD3 model in the drones selects an energy-efficient path without the overhead of the training process of the RL model in real-time, considering external factors such as wind or another drone. The total energy consumption of drones in flight along with online path planning is approximately 106% of the total energy consumption of drones that follow offline path planning, even though the trained TD3 model does not require complex computations for real-time execution.-
dc.languageEnglish-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleEnergy-Efficient Online Path Planning of Multiple Drones Using Reinforcement Learning-
dc.typeArticle-
dc.identifier.wosid000707443200011-
dc.identifier.scopusid2-s2.0-85112222850-
dc.type.rimsART-
dc.citation.volume70-
dc.citation.issue10-
dc.citation.beginningpage9725-
dc.citation.endingpage9740-
dc.citation.publicationnameIEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY-
dc.identifier.doi10.1109/TVT.2021.3102589-
dc.embargo.liftdate9999-12-31-
dc.embargo.terms9999-12-31-
dc.contributor.localauthorChang, Naehyuck-
dc.contributor.nonIdAuthorBaek, Donkyu-
dc.contributor.nonIdAuthorKim, Jaemin-
dc.description.isOpenAccessN-
dc.type.journalArticleArticle-
dc.subject.keywordAuthorDrones-
dc.subject.keywordAuthorPath planning-
dc.subject.keywordAuthorConvergence-
dc.subject.keywordAuthorReal-time systems-
dc.subject.keywordAuthorComputational modeling-
dc.subject.keywordAuthorTraining-
dc.subject.keywordAuthorInterference-
dc.subject.keywordAuthorMulti-robot systems-
dc.subject.keywordAuthorcyber-physical systems-
dc.subject.keywordAuthorunmanned aerial vehicles-
dc.subject.keywordAuthorpath planning-
dc.subject.keywordAuthorreinforcement learning-
Appears in Collection
EE-Journal Papers(저널논문)
Files in This Item
This item is cited by other documents in WoS
⊙ Detail Information in WoSⓡ Click to see webofscience_button
⊙ Cited 32 items in WoS Click to see citing articles in records_button

qr_code

  • mendeley

    citeulike


rss_1.0 rss_2.0 atom_1.0