Convolutional Neural Network-Based Vehicle Classification in Low-Quality Imaging Conditions for Internet of Things Devices
| dc.contributor.author | Maiga,B. | |
| dc.contributor.author | Dalveren,Y. | |
| dc.contributor.author | Kara,A. | |
| dc.contributor.author | Derawi,M. | |
| dc.contributor.other | Department of Electrical & Electronics Engineering | |
| dc.contributor.other | 15. Graduate School of Natural and Applied Sciences | |
| dc.contributor.other | 01. Atılım University | |
| dc.date.accessioned | 2024-07-05T15:23:27Z | |
| dc.date.available | 2024-07-05T15:23:27Z | |
| dc.date.issued | 2023 | |
| dc.description | Kara, Ali/0000-0002-9739-7619; Dalveren, Yaser/0000-0002-9459-0042 | en_US |
| dc.description.abstract | Vehicle classification has an important role in the efficient implementation of Internet of Things (IoT)-based intelligent transportation system (ITS) applications. Nowadays, because of their higher performance, convolutional neural networks (CNNs) are mostly used for vehicle classification. However, the computational complexity of CNNs and high-resolution data provided by high-quality monitoring cameras can pose significant challenges due to limited IoT device resources. In order to address this issue, this study aims to propose a simple CNN-based model for vehicle classification in low-quality images collected by a standard security camera positioned far from a traffic scene under low lighting and different weather conditions. For this purpose, firstly, a new dataset that contains 4800 low-quality vehicle images with 100 × 100 pixels and a 96 dpi resolution was created. Then, the proposed model and several well-known CNN-based models were tested on the created dataset. The results demonstrate that the proposed model achieved 95.8% accuracy, outperforming Inception v3, Inception-ResNet v2, Xception, and VGG19. While DenseNet121 and ResNet50 achieved better accuracy, their complexity in terms of higher trainable parameters, layers, and training times might be a significant concern in practice. In this context, the results suggest that the proposed model could be a feasible option for IoT devices used in ITS applications due to its simple architecture. © 2023 by the authors. | en_US |
| dc.identifier.doi | 10.3390/su152316292 | |
| dc.identifier.issn | 2071-1050 | |
| dc.identifier.scopus | 2-s2.0-85188720986 | |
| dc.identifier.uri | https://doi.org/10.3390/su152316292 | |
| dc.language.iso | en | en_US |
| dc.publisher | Multidisciplinary Digital Publishing Institute (MDPI) | en_US |
| dc.relation.ispartof | Sustainability (Switzerland) | en_US |
| dc.rights | info:eu-repo/semantics/openAccess | en_US |
| dc.subject | bad weather | en_US |
| dc.subject | deep learning | en_US |
| dc.subject | intelligent transportation system | en_US |
| dc.subject | tiny images | en_US |
| dc.title | Convolutional Neural Network-Based Vehicle Classification in Low-Quality Imaging Conditions for Internet of Things Devices | en_US |
| dc.type | Article | en_US |
| dspace.entity.type | Publication | |
| gdc.author.id | Kara, Ali/0000-0002-9739-7619 | |
| gdc.author.id | Dalveren, Yaser/0000-0002-9459-0042 | |
| gdc.author.institutional | Dalveren, Yaser | |
| gdc.author.institutional | Kara, Ali | |
| gdc.author.scopusid | 57222965218 | |
| gdc.author.scopusid | 51763497600 | |
| gdc.author.scopusid | 7102824862 | |
| gdc.author.scopusid | 35408917600 | |
| gdc.bip.impulseclass | C5 | |
| gdc.bip.influenceclass | C5 | |
| gdc.bip.popularityclass | C4 | |
| gdc.coar.access | open access | |
| gdc.coar.type | text::journal::journal article | |
| gdc.description.department | Atılım University | en_US |
| gdc.description.departmenttemp | Maiga B., Graduate School of Natural and Applied Sciences, Department of Electrical and Electronics Engineering, Atilim University, Ankara, 06830, Turkey; Dalveren Y., Department of Electrical and Electronics Engineering, Atilim University, Ankara, 06830, Turkey; Kara A., Department of Electrical and Electronics Engineering, Gazi University, Ankara, 06570, Turkey; Derawi M., Department of Electronic Systems, Norwegian University of Science and Technology, Gjovik, 2815, Norway | en_US |
| gdc.description.issue | 23 | en_US |
| gdc.description.publicationcategory | Makale - Uluslararası Hakemli Dergi - Kurum Öğretim Elemanı | en_US |
| gdc.description.scopusquality | Q2 | |
| gdc.description.startpage | 16292 | |
| gdc.description.volume | 15 | en_US |
| gdc.description.wosquality | Q2 | |
| gdc.identifier.openalex | W4388968971 | |
| gdc.identifier.wos | WOS:001116187600001 | |
| gdc.oaire.accesstype | GOLD | |
| gdc.oaire.diamondjournal | false | |
| gdc.oaire.impulse | 4.0 | |
| gdc.oaire.influence | 2.8795808E-9 | |
| gdc.oaire.isgreen | false | |
| gdc.oaire.popularity | 5.438212E-9 | |
| gdc.oaire.publicfunded | false | |
| gdc.oaire.sciencefields | 0202 electrical engineering, electronic engineering, information engineering | |
| gdc.oaire.sciencefields | 02 engineering and technology | |
| gdc.oaire.sciencefields | 01 natural sciences | |
| gdc.oaire.sciencefields | 0104 chemical sciences | |
| gdc.openalex.fwci | 0.511 | |
| gdc.openalex.normalizedpercentile | 0.69 | |
| gdc.opencitations.count | 0 | |
| gdc.plumx.mendeley | 18 | |
| gdc.plumx.newscount | 1 | |
| gdc.plumx.scopuscites | 5 | |
| gdc.scopus.citedcount | 5 | |
| gdc.wos.citedcount | 3 | |
| relation.isAuthorOfPublication | 55e082ac-14c0-46a6-b8fa-50c5e40b59c8 | |
| relation.isAuthorOfPublication | be728837-c599-49c1-8e8d-81b90219bb15 | |
| relation.isAuthorOfPublication.latestForDiscovery | 55e082ac-14c0-46a6-b8fa-50c5e40b59c8 | |
| relation.isOrgUnitOfPublication | c3c9b34a-b165-4cd6-8959-dc25e91e206b | |
| relation.isOrgUnitOfPublication | dff2e5a6-d02d-4bef-8b9e-efebe3919b10 | |
| relation.isOrgUnitOfPublication | 50be38c5-40c4-4d5f-b8e6-463e9514c6dd | |
| relation.isOrgUnitOfPublication.latestForDiscovery | c3c9b34a-b165-4cd6-8959-dc25e91e206b |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- Convolutional Neural Network-Based sustainability-15-16292-v2.pdf
- Size:
- 2.12 MB
- Format:
- Adobe Portable Document Format