A new artificial intelligence framework can create a highly accurate digital twin of industrial distillation columns, predicting internal chemical compositions and temperatures in real time while strictly adhering to thermodynamic laws. This breakthrough addresses a critical in chemical engineering, where distillation accounts for roughly 40% of industrial energy consumption globally, yet operators often lack detailed measurements due to the high cost and maintenance difficulties of sensors on every tray. By combining machine learning with fundamental physics, the model offers a practical solution for monitoring and optimizing these energy-intensive processes without relying solely on expensive instrumentation or slow, computationally heavy simulations.
The researchers developed a Physics-Informed Neural Network (PINN) that achieves a root mean square error (RMSE) of 0.00143 for predicting the mole fraction of a key chemical component, HX, with an R-squared value of 0.9887. This represents a 44.6% reduction in RMSE compared to the best data-driven baseline model, a Transformer, while ensuring all predictions satisfy thermodynamic constraints like vapor-liquid equilibrium and mass balance. The model was trained and evaluated on a synthetic dataset of 961 timestamped measurements spanning eight hours of transient operation, generated using Aspen HYSYS software for a binary distillation system separating HX and TX mixtures. It accurately captures dynamic responses to perturbations such as changes in reflux ratio, feed flow rates, and pressure, reconstructing tray-wise temperature and composition profiles that align with physical expectations.
Ology embeds core thermodynamic principles directly into the neural network's training process through a composite loss function. This includes constraints from modified Raoult's law for vapor-liquid equilibrium, tray-level mass and energy balances (known as MESH equations), and the McCabe-Thiele graphical for distillation. The PINN architecture uses a fully connected network with four hidden layers and Swish activation functions, taking 16 sensor inputs—such as condenser pressure, reboiler liquid hold-up, and reflux ratio—alongside time data to predict outputs including mole fractions and tray temperatures. An adaptive weighting scheme balances data fidelity and physics consistency during training, starting with a strong emphasis on physical laws before gradually shifting focus to matching the measured data, which prevents early over-regularization and improves generalization.
From the study, detailed in figures throughout the paper, show the PINN's superior performance across multiple metrics. For instance, Figure 7 illustrates that the PINN closely tracks HX and TX mole fractions over time, with a parity plot indicating tight clustering around the ideal prediction line. Table 2 compares the PINN against five data-driven baselines—LSTM, vanilla MLP, GRU, Transformer, and DeepONet—revealing that the PINN not only has the lowest MSE (2.05×10⁻⁶) and RMSE but is also the only model to consistently satisfy thermodynamic constraints, with a mean VLE residual of 1.8×10⁻⁵ versus 3.4×10⁻³ for the Transformer. Additionally, Figure 8 demonstrates the model's ability to reconstruct physically plausible tray-wise profiles, such as S-shaped composition curves that sharpen with increasing reflux, details that data-only models often miss.
Of this research are significant for industrial applications, particularly in real-time soft sensing, model-predictive control, and anomaly detection in distillation processes. By providing accurate, thermodynamically consistent predictions of unmeasured variables, the digital twin can help operators optimize energy use and maintain product purity without the need for extensive sensor arrays. The adaptive weighting approach also proves effective in low-data regimes, as shown in Figure 10, where the PINN maintains an RMSE of 0.00198 with only 30% of the training data, compared to 0.00441 for the Transformer, suggesting robustness in scenarios with limited measurements. This makes the framework a promising foundation for broader use in chemical process modeling, where physical laws are critical but data may be scarce or noisy.
However, the study acknowledges several limitations that must be addressed before industrial deployment. The training and evaluation were conducted on a synthetic dataset with known ground truth, which avoids real-world s like sensor drift, measurement dropouts, and calibration uncertainty. Future work plans to validate the model on real plant historian data to assess its performance under these conditions. Additionally, the current implementation assumes simplifications such as equimolar overflow in the McCabe-Thiele constraint, which may not hold for systems with highly non-ideal vapor-liquid equilibrium; the researchers intend to replace this with more rigorous equations in future revisions. Extensions to multi-component systems, reactive distillation, and integration with Bayesian s for uncertainty quantification are also noted as important directions for further research.
Original Source
Read the complete research paper
About the Author
Guilherme A.
Former dentist (MD) from Brazil, 41 years old, husband, and AI enthusiast. In 2020, he transitioned from a decade-long career in dentistry to pursue his passion for technology, entrepreneurship, and helping others grow.
Connect on LinkedIn