A reinforcement learning framework for improving parking decisions in last-mile delivery

被引:1
|
作者
Muriel, Juan E. [1 ,2 ]
Zhang, Lele [3 ,4 ]
Fransoo, Jan C. [5 ]
Villegas, Juan G. [6 ]
机构
[1] Pacific Natl, Brisbane, Qld 4006, Australia
[2] Deakin Univ, Sch Engn & Built Environm SEBE, Geelong, Vic 3216, Australia
[3] Univ Melbourne, Sch Math & Stat, Parkville, Australia
[4] ARC Training Ctr Optimisat Technol Integrated Meth, Clayton, Australia
[5] Tilburg Univ, Tilburg Sch Econ & Management, Tilburg, Netherlands
[6] Univ Antioquia, Fac Ingn, Dept Ingn Ind, Medellin, Colombia
关键词
Last-mile delivery; urban logistics; reinforcement learning; loading zone; simulation-optimisation; URBAN FREIGHT TRANSPORT; CITY; AREAS; SYSTEMS; SIMULATION; LOCATION; VEHICLES; IMPACT; TOOLS; BAYS;
D O I
10.1080/21680566.2024.2337216
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
This study leverages simulation-optimisation with a Reinforcement Learning (RL) model to analyse the routing behaviour of delivery vehicles (DVs). We conceptualise the system as a stochastic k-armed bandit problem, representing a sequential interaction between a learner (the DV) and its surrounding environment. Each DV is assigned a random number of customers and an initial delivery route. If a loading zone is unavailable, the RL model is used to select a delivery strategy, thereby modifying its route accordingly. The penalty is gauged by the additional trucking and walking time incurred compared to the originally planned route. Our methodology is tested on a simulated network featuring realistic traffic conditions and a fleet of DVs employing four distinct lastmile delivery strategies. The results of our numerical experiments underscore the advantages of providing DVs with an RL-based decision support system for en-route decision-making, yielding benefits to the overall efficiency of the transport network.HighlightsCombining simulation and optimisation algorithms with reinforcement learningModel DVs en-route parking decisions with a k-armed bandit algorithmEvaluating the impacts of delivery strategies on traffic congestion and in last-mile delivery efficiency
引用
收藏
页数:29
相关论文
共 50 条
  • [1] Deep reinforcement learning for stochastic last-mile delivery with crowdshipping
    Silva, Marco
    Pedroso, Joao Pedro
    Viana, Ana
    EURO JOURNAL ON TRANSPORTATION AND LOGISTICS, 2023, 12
  • [2] Parking Guidance and Geofencing for Last-Mile Delivery Operations
    Simoni, Michele D.
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (08) : 9091 - 9102
  • [3] A contextual framework for learning routing experiences in last-mile delivery
    Sun, Huai Jun
    Arslan, Okan
    TRANSPORTATION RESEARCH PART B-METHODOLOGICAL, 2025, 194
  • [4] Deep Reinforcement Learning for Crowdshipping Last-Mile Delivery with Endogenous Uncertainty
    Silva, Marco
    Pedroso, Joao Pedro
    MATHEMATICS, 2022, 10 (20)
  • [5] On the Regulatory Framework for Last-Mile Delivery Robots
    Hoffmann, Thomas
    Prause, Gunnar
    MACHINES, 2018, 6 (03)
  • [6] Does parking matter? The impact of parking time on last-mile delivery optimization
    Reed, Sara
    Campbell, Ann Melissa
    Thomas, Barrett W.
    TRANSPORTATION RESEARCH PART E-LOGISTICS AND TRANSPORTATION REVIEW, 2024, 181
  • [7] A Bilevel Approach for Compensation and Routing Decisions in Last-Mile Delivery
    Cerulli, Martina
    Archetti, Claudia
    Fernandez, Elena
    Ljubic, Ivana
    TRANSPORTATION SCIENCE, 2024, 58 (05) : 1076 - 1100
  • [8] Tactical workforce sizing and scheduling decisions for last-mile delivery
    Mandal, Minakshi Punam
    Santini, Alberto
    Archetti, Claudia
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2025, 323 (01) : 153 - 169
  • [9] Crowdkeeping in Last-Mile Delivery
    Wang, Xin
    Arslan, Okan
    Delage, Erick
    TRANSPORTATION SCIENCE, 2024, 58 (02) : 474 - 498
  • [10] Recent applications for improving the last-mile delivery in urbanism logistics
    Bui, Viet Duc
    Nguyen, Hoang Phuong
    Nguyen, Thi Tuyet Mai
    INTERNATIONAL JOURNAL OF KNOWLEDGE-BASED DEVELOPMENT, 2022, 12 (3-4) : 328 - 346