A Reinforcement Learning Algorithm for Data Collection in Uav-Aided Iot Networks With Uncertain Time Windows

dc.authorid Cicek, Cihan Tugrul/0000-0002-3532-2638
dc.authorscopusid 57208147005
dc.authorwosid Cicek, Cihan Tugrul/AAF-7787-2019
dc.contributor.author Cicek, Cihan Tugrul
dc.contributor.other Industrial Engineering
dc.date.accessioned 2024-07-05T15:19:11Z
dc.date.available 2024-07-05T15:19:11Z
dc.date.issued 2021
dc.department Atılım University en_US
dc.department-temp [Cicek, Cihan Tugrul] Atilim Univ, Dept Ind Engn, Ankara, Turkey en_US
dc.description Cicek, Cihan Tugrul/0000-0002-3532-2638 en_US
dc.description.abstract Unmanned aerial vehicles (UAVs) have been considered as an efficient solution to collect data from ground sensor nodes in Internet-of-Things (IoT) networks due to their several advantages such as flexibility, quick deployment and maneuverability. Studies on this subject have been mainly focused on problems where limited UAV battery is introduced as a tight constraint that shortens the mission time in the models, which significantly undervalues the UAV potential. Moreover, the sensors in the network are typically assumed to have deterministic working times during which the data is uploaded. In this study, we revisit the UAV trajectory planning problem with a different approach and revise the battery constraint by allowing UAVs to swap their batteries at fixed stations and continue their data collection task, hence, the planning horizon can be extended. In particular, we develop a discrete time Markov process (DTMP) in which the UAV trajectory and battery swapping times are jointly determined to minimize the total data loss in the network, where the sensors have uncertain time windows for uploading. Due to the so-called curse-of-dimensionality, we propose a reinforcement learning (RL) algorithm in which the UAV is trained as an agent to explore the network. The computational study shows that our proposed algorithm outperforms two benchmark approaches and achieves significant reduction in data loss. en_US
dc.identifier.citationcount 0
dc.identifier.doi 10.1109/ICCWorkshops50388.2021.9473768
dc.identifier.isbn 9781728194417
dc.identifier.issn 2164-7038
dc.identifier.scopus 2-s2.0-85112795751
dc.identifier.uri https://doi.org/10.1109/ICCWorkshops50388.2021.9473768
dc.identifier.uri https://hdl.handle.net/20.500.14411/1947
dc.identifier.wos WOS:000848412200244
dc.institutionauthor Çiçek, Cihan Tuğrul
dc.language.iso en en_US
dc.publisher Ieee en_US
dc.relation.ispartof IEEE International Conference on Communications (ICC) -- JUN 14-23, 2021 -- ELECTR NETWORK en_US
dc.relation.ispartofseries IEEE International Conference on Communications Workshops
dc.relation.publicationcategory Konferans Öğesi - Uluslararası - Kurum Öğretim Elemanı en_US
dc.rights info:eu-repo/semantics/closedAccess en_US
dc.scopus.citedbyCount 6
dc.subject UAV en_US
dc.subject internet-of-things en_US
dc.subject reinforcement learning en_US
dc.subject battery swapping en_US
dc.subject time windows en_US
dc.subject uncertainty en_US
dc.title A Reinforcement Learning Algorithm for Data Collection in Uav-Aided Iot Networks With Uncertain Time Windows en_US
dc.type Conference Object en_US
dc.wos.citedbyCount 3
dspace.entity.type Publication
relation.isAuthorOfPublication 82ea98fd-36fb-4469-8e29-b73dc71cabb9
relation.isAuthorOfPublication.latestForDiscovery 82ea98fd-36fb-4469-8e29-b73dc71cabb9
relation.isOrgUnitOfPublication 12c9377e-b7fe-4600-8326-f3613a05653d
relation.isOrgUnitOfPublication.latestForDiscovery 12c9377e-b7fe-4600-8326-f3613a05653d

Files

Collections