Abstract

The gasoline inline blending process has widely used real-time optimization techniques to achieve optimization objectives, such as minimizing the cost of production. However, the effectiveness of real-time optimization in gasoline blending relies on accurate blending models and is challenged by stochastic disturbances. Thus, we propose a real-time optimization algorithm based on the Soft Actor-Critic deep reinforcement learning strategy to optimize gasoline blending without relying on a single blending model and to be robust against disturbances. Our approach constructs the environment using nonlinear blending models and feedstocks with disturbances. The algorithm incorporates the Lagrange multiplier and path constraints in reward design to manage sparse product constraints. Carefully abstracted states facilitate algorithm convergence, and the normalized action vector in each optimization period allows the agent to generalize to some extent across different target production scenarios. Through these well-designed components, the algorithm based on the Soft Actor-Critic outperforms real-time optimization methods based on either nonlinear or linear programming. It even demonstrates comparable performance with the time-horizon based real-time optimization method, which requires knowledge of uncertainty models, confirming its capability to handle uncertainty without accurate models. Our simulation illustrates a promising approach to free real-time optimization of the gasoline blending process from uncertainty models that are difficult to acquire in practice.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call