Deep Reinforcement Learning Algorithm for Dynamic Pricing of Express Lanes with Multiple Access Locations
September 10, 2019 ยท Entered Twilight ยท ๐ Transportation Research Part C: Emerging Technologies
"Last commit was 5.0 years ago (โฅ5 year threshold)"
Evidence collected by the PWNC Scanner
Repo contents: .DS_Store, .Rhistory, LICENSE, README.md, __init__.py, __pycache__, data, gym-meme, requirements.txt, setup.py, spinup
Authors
Venktesh Pandey, Evana Wang, Stephen D. Boyles
arXiv ID
1909.04760
Category
eess.SY: Systems & Control (EE)
Cross-listed
cs.AI,
cs.LG,
stat.ML
Citations
43
Venue
Transportation Research Part C: Emerging Technologies
Repository
https://github.com/venktesh22/ExpressLanes_Deep-RL
โญ 26
Last Checked
1 month ago
Abstract
This article develops a deep reinforcement learning (Deep-RL) framework for dynamic pricing on managed lanes with multiple access locations and heterogeneity in travelers' value of time, origin, and destination. This framework relaxes assumptions in the literature by considering multiple origins and destinations, multiple access locations to the managed lane, en route diversion of travelers, partial observability of the sensor readings, and stochastic demand and observations. The problem is formulated as a partially observable Markov decision process (POMDP) and policy gradient methods are used to determine tolls as a function of real-time observations. Tolls are modeled as continuous and stochastic variables, and are determined using a feedforward neural network. The method is compared against a feedback control method used for dynamic pricing. We show that Deep-RL is effective in learning toll policies for maximizing revenue, minimizing total system travel time, and other joint weighted objectives, when tested on real-world transportation networks. The Deep-RL toll policies outperform the feedback control heuristic for the revenue maximization objective by generating revenues up to 9.5% higher than the heuristic and for the objective minimizing total system travel time (TSTT) by generating TSTT up to 10.4% lower than the heuristic. We also propose reward shaping methods for the POMDP to overcome the undesired behavior of toll policies, like the jam-and-harvest behavior of revenue-maximizing policies. Additionally, we test transferability of the algorithm trained on one set of inputs for new input distributions and offer recommendations on real-time implementations of Deep-RL algorithms. The source code for our experiments is available online at https://github.com/venktesh22/ExpressLanes_Deep-RL
Community Contributions
Found the code? Know the venue? Think something is wrong? Let us know!
๐ Similar Papers
In the same crypt โ Systems & Control (EE)
R.I.P.
๐ป
Ghosted
R.I.P.
๐ป
Ghosted
Incremental Gradient, Subgradient, and Proximal Methods for Convex Optimization: A Survey
R.I.P.
๐ป
Ghosted
Wireless Network Design for Control Systems: A Survey
R.I.P.
๐ป
Ghosted
Learning-based Model Predictive Control for Safe Exploration
R.I.P.
๐ป
Ghosted
Safety-Critical Model Predictive Control with Discrete-Time Control Barrier Function
R.I.P.
๐ป
Ghosted