APS Logo

Control of Partially Observable Flows with Model-Free Reinforcement Learning

ORAL

Abstract

Model-free Reinforcement Learning algorithms have been employed recently to discover flow control strategies, for example for efficient drag reduction by suppressing vortex-shedding in the wake of a circular cylinder. These methods rely on probes located in the flow downstream of the body to achieve full-state observability and control. The present approach considers real-world applicability by restricting sensing to pressure probes mounted on the base of a square bluff body. Surface mounted sensing is shown to restrict observability over the flow and reduce drag reduction performance by 65% compared to probes optimally located downstream of the body. A method integrating memory into the control architecture is proposed to improve drag reduction performance in partially observable systems. Memory is integrated by augmenting the input to the controller with a time series of lagged observations. A power expenditure study shows that the active drag reduction strategies discovered with Reinforcement Learning are extremely power efficient. These results are a first step towards realistic implementation of reinforcement learning for active drag reduction in the type of partially observable systems often found in the real-world.

Presenters

  • Georgios Rigas

    Imperial College London

Authors

  • Georgios Rigas

    Imperial College London