dc.contributor.author | Konovalenko, Anna | |
dc.contributor.author | Hvattum, Lars Magnus | |
dc.date.accessioned | 2024-10-07T09:17:19Z | |
dc.date.available | 2024-10-07T09:17:19Z | |
dc.date.created | 2024-10-02T11:56:04Z | |
dc.date.issued | 2024 | |
dc.identifier.citation | Logistics. 2024, 8 (4), 96 | en_US |
dc.identifier.issn | 2305-6290 | |
dc.identifier.uri | https://hdl.handle.net/11250/3156615 | |
dc.description.abstract | Background: The dynamic vehicle routing problem (DVRP) is a complex optimization problem that is crucial for applications such as last-mile delivery. Our goal is to develop an application that can make real-time decisions to maximize total performance while adapting to the dynamic nature of incoming orders. We formulate the DVRP as a vehicle routing problem where new customer requests arrive dynamically, requiring immediate acceptance or rejection decisions. Methods: This study leverages reinforcement learning (RL), a machine learning paradigm that operates via feedback-driven decisions, to tackle the DVRP. We present a detailed RL formulation and systematically investigate the impacts of various state-space components on algorithm performance. Our approach involves incrementally modifying the state space, including analyzing the impacts of individual components, applying data transformation methods, and incorporating derived features. Results: Our findings demonstrate that a carefully designed state space in the formulation of the DVRP significantly improves RL performance. Notably, incorporating derived features and selectively applying feature transformation enhanced the model’s decision-making capabilities. The combination of all enhancements led to a statistically significant improvement in the results compared with the basic state formulation. Conclusions: This research provides insights into RL modeling for DVRPs, highlighting the importance of state-space design. The proposed approach offers a flexible framework that is applicable to various variants of the DVRP, with potential for validation using real-world data. | en_US |
dc.language.iso | eng | en_US |
dc.rights | Navngivelse 4.0 Internasjonal | * |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/deed.no | * |
dc.title | Optimizing a Dynamic Vehicle Routing Problem with Deep Reinforcement Learning: Analyzing State-Space Components | en_US |
dc.type | Peer reviewed | en_US |
dc.type | Journal article | en_US |
dc.description.version | publishedVersion | en_US |
dc.source.volume | 8 | en_US |
dc.source.journal | Logistics | en_US |
dc.source.issue | 4 | en_US |
dc.identifier.doi | 10.3390/logistics8040096 | |
dc.identifier.cristin | 2308684 | |
dc.source.articlenumber | 96 | en_US |
cristin.ispublished | true | |
cristin.fulltext | original | |
cristin.qualitycode | 1 | |