Learn to Flap: Foil Non-parametric Path Planning via Deep Reinforcement Learning

To optimize flapping foil performance, the application of deep reinforcement learning (DRL) on controlling foil non-parametric motion is conducted in the present study. Traditional control techniques and simplified motions cannot fully model nonlinear, unsteady and high-dimensional foil-vortex inter...

Full description

Saved in:
Bibliographic Details
Published inarXiv.org
Main Authors Wang, Z P, Lin, R J, Zhao, Z Y, Guo, P M, Yang, N, Fan, D X
Format Paper
LanguageEnglish
Published Ithaca Cornell University Library, arXiv.org 25.05.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:To optimize flapping foil performance, the application of deep reinforcement learning (DRL) on controlling foil non-parametric motion is conducted in the present study. Traditional control techniques and simplified motions cannot fully model nonlinear, unsteady and high-dimensional foil-vortex interactions. A DRL-training framework based on Proximal Policy Optimization and Transformer architecture is proposed. The policy is initialized from the sinusoidal expert display. We first demonstrate the effectiveness of the proposed DRL-training framework which can optimize foil motion while enhancing foil generated thrust. By adjusting reward setting and action threshold, the DRL-optimized foil trajectories can gain further enhancement compared to sinusoidal motion. Via flow analysis of wake morphology and instantaneous pressure distributions, it is found that the DRL-optimized foil can adaptively adjust the phases between motion and shedding vortices to improve hydrodynamic performance. Our results give a hint for solving complex fluid manipulation problems through DRL method.
ISSN:2331-8422