Model Predictive Control has been recently proposed as policy approximation for Reinforcement Learning, offering a path towards safe and explainable Reinforcement Learning. This approach has been investigated for Q-learning and actor-critic methods, both in the context of nominal Economic MPC and Robust (N)MPC, showing very promising results. In that context, actor-critic methods seem to be the most reliable approach. Many applications include a mixture of continuous and integer inputs, for which the classical actor-critic methods need to be adapted. In this paper, we present a policy approximation based on mixed-integer MPC schemes, and propose a computationally inexpensive technique to generate exploration in the mixed-integer input space that ensures a satisfaction of the constraints. We then propose a simple compatible advantage function approximation for the proposed policy, that allows one to build the gradient of the mixed-integer MPC-based policy.


    Zugriff

    Download


    Exportieren, teilen und zitieren



    Titel :

    Reinforcement learning for mixed-integer problems based on MPC



    Erscheinungsdatum :

    2020-01-01



    Medientyp :

    Aufsatz (Konferenz)


    Format :

    Elektronische Ressource


    Sprache :

    Englisch



    Klassifikation :

    DDC:    629



    Learning contact-aware robot controllers from mixed integer optimization

    Deits, Robin L. H.(Robin Lloyd Henderson) | BASE | 2019

    Freier Zugriff




    Hierarchical Reinforcement Learning Under Mixed Observability

    Nguyen, Hai / Yang, Zhihan / Baisero, Andrea et al. | Springer Verlag | 2022