Reinforcement Learning Using Fully Connected, Attention, and Transformer Models in Knapsack Problem Solving

dc.authorid YILDIZ, Beytullah/0000-0001-7664-5145
dc.authorscopusid 14632851900
dc.contributor.author Yildiz, Beytullah
dc.contributor.author Yıldız, Beytullah
dc.contributor.author Yıldız, Beytullah
dc.contributor.other Software Engineering
dc.date.accessioned 2024-07-05T15:21:19Z
dc.date.available 2024-07-05T15:21:19Z
dc.date.issued 2022
dc.department Atılım University en_US
dc.department-temp [Yildiz, Beytullah] Atilim Univ, Sch Engn, Dept Software Engn, Ankara, Turkey en_US
dc.description YILDIZ, Beytullah/0000-0001-7664-5145 en_US
dc.description.abstract Knapsack is a combinatorial optimization problem that involves a variety of resource allocation challenges. It is defined as non-deterministic polynomial time (NP) hard and has a wide range of applications. Knapsack problem (KP) has been studied in applied mathematics and computer science for decades. Many algorithms that can be classified as exact or approximate solutions have been proposed. Under the category of exact solutions, algorithms such as branch-and-bound and dynamic programming and the approaches obtained by combining these algorithms can be classified. Due to the fact that exact solutions require a long processing time, many approximate methods have been introduced for knapsack solution. In this research, deep Q-learning using models containing fully connected layers, attention, and transformer as function estimators were used to provide the solution for KP. We observed that deep Q-networks, which continued their training by observing the reward signals provided by the knapsack environment we developed, optimized the total reward gained over time. The results showed that our approaches give near-optimum solutions and work about 40 times faster than an exact algorithm using dynamic programming. en_US
dc.identifier.citationcount 1
dc.identifier.doi 10.1002/cpe.6509
dc.identifier.issn 1532-0626
dc.identifier.issn 1532-0634
dc.identifier.issue 9 en_US
dc.identifier.scopus 2-s2.0-85112660920
dc.identifier.scopusquality Q2
dc.identifier.uri https://doi.org/10.1002/cpe.6509
dc.identifier.uri https://hdl.handle.net/20.500.14411/2049
dc.identifier.volume 34 en_US
dc.identifier.wos WOS:000682738000001
dc.identifier.wosquality Q3
dc.institutionauthor Yıldız, Beytullah
dc.language.iso en en_US
dc.publisher Wiley en_US
dc.relation.publicationcategory Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı en_US
dc.rights info:eu-repo/semantics/closedAccess en_US
dc.scopus.citedbyCount 18
dc.subject attention en_US
dc.subject combinatorial optimization problem en_US
dc.subject deep Q-learning en_US
dc.subject knapsack en_US
dc.subject reinforcement learning en_US
dc.subject transformer en_US
dc.title Reinforcement Learning Using Fully Connected, Attention, and Transformer Models in Knapsack Problem Solving en_US
dc.type Article en_US
dc.wos.citedbyCount 8
dspace.entity.type Publication
relation.isAuthorOfPublication 8eb144cb-95ff-4557-a99c-cd0ffa90749d
relation.isAuthorOfPublication 8eb144cb-95ff-4557-a99c-cd0ffa90749d
relation.isAuthorOfPublication 8eb144cb-95ff-4557-a99c-cd0ffa90749d
relation.isAuthorOfPublication.latestForDiscovery 8eb144cb-95ff-4557-a99c-cd0ffa90749d
relation.isOrgUnitOfPublication d86bbe4b-0f69-4303-a6de-c7ec0c515da5
relation.isOrgUnitOfPublication.latestForDiscovery d86bbe4b-0f69-4303-a6de-c7ec0c515da5

Files

Collections