To effectively solve Demand and Capacity Balancing (DCB) in large-scale and high-density scenarios through the Ground Delay Program (GDP) in the pre-tactical stage, a sequential decision-making framework based on a time window is proposed. On this basis, the problem is transformed into Markov Decision Process (MDP) based on local observation, and then Multi-Agent Reinforcement Learning (MARL) method is adopted. Each flight is regarded as an independent agent to decide whether to implement GDP according to its local state observation. By designing the reward function in multiple combinations, a Mixed Competition and Cooperation (MCC) mode considering fairness is formed among agents. To improve the efficiency of MARL, we use the double Q-Learning Network (DQN), experience replay technology, adaptive ϵ-greedy strategy and Decentralized Training with Decentralized Execution (DTDE) framework. The experimental results show that the training process of the MARL method is convergent, efficient and stable. Compared with the Computer-Assisted Slot Allocation (CASA) method used in the actual operation, the number of flight delays and the average delay time is reduced by 33.7% and 36.7% respectively.


    Access

    Check access

    Check availability in my library

    Order at Subito €


    Export, share and cite



    Title :

    Demand and Capacity Balancing Technology Based on Multi-agent Reinforcement Learning


    Contributors:
    Chen, Yutong (author) / Xu, Yan (author) / Hu, Minghua (author) / Yang, Lei (author)


    Publication date :

    2021-10-03


    Size :

    2106886 byte





    Type of media :

    Conference paper


    Type of material :

    Electronic Resource


    Language :

    English




    Multi-Agent Reinforcement Learning for Autonomous On Demand Vehicles

    Boyali, Ali / Hashimoto, Naohisa / John, Vijay et al. | IEEE | 2019



    Multiagent Reinforcement Learning Methods for Resolving Demand - Capacity Imbalances

    Kravaris, Theocharis / Spatharis, Christos / Blekas, Konstantions et al. | IEEE | 2018