Off-policy Q-learning: set-point design for optimizing dual-rate rougher flotation operational processes

J Li, T Y Chai, F Lewis, J. Fan, Zhengtao Ding, J Ding

    Research output: Contribution to journalArticlepeer-review

    291 Downloads (Pure)

    Abstract

    Rougher flotation, composed of unit processes operating at a fast time scale and economic performance measurements known as operational indices measured at a slower time scale, is very basic and the first concentration stage for flotation plants. Optimizing operational process for rougher flotation circuits is extremely important due to high economic profit arising from the optimality of operational indices. This paper presents a novel off-policy Q-learning method to learn the
    optimal solution to rougher flotation operational processes without the knowledge of dynamics of unit processes and operational indices. To this end, first, the optimal operational control (OOC) for dual-rate rougher flotation
    processes is formulated. Second, H∞ tracking control problem is developed to optimally prescribe the set-points for the rougher flotation processes. Then, a zero-sum game off-policy Q-learning algorithm is proposed to find the
    optimal set-points by using measured data. Finally, simulation experiments are employed to show the effectiveness of the proposed method.
    Original languageEnglish
    JournalIEEE Transactions on Industrial Electronics
    DOIs
    Publication statusPublished - 6 Oct 2017

    Keywords

    • Rougher flotation
    • Operational optimization
    • Q-learning
    • Zero-sum game
    • H∞ tracking control

    Fingerprint

    Dive into the research topics of 'Off-policy Q-learning: set-point design for optimizing dual-rate rougher flotation operational processes'. Together they form a unique fingerprint.

    Cite this