In this paper, negatively inclined buoyant jets, which appear during the discharge of wastewater from processes such as desalination, are observed. To minimize harmful effects and assess environmental impact, a detailed numerical investigation is necessary. The selection of appropriate geometry and working conditions for minimizing such effects often requires numerous experiments and numerical simulations. For this reason, the application of machine learning models is proposed. Several models including Support Vector Regression, Artificial Neural Networks, Random Forests, XGBoost, CatBoost and LightGBM were trained. The dataset was built with numerous OpenFOAM simulations, which were validated by experimental data from previous research. The best prediction was obtained by Artificial Neural Network with an average of R2 0.98 and RMSE 0.28. In order to understand the working of the machine learning model and the influence of all parameters on the geometrical characteristics of inclined buoyant jets, the SHAP feature interpretation method was used.
translated by 谷歌翻译
在本文中,使用计算流体动力学研究了具有次级通道和肋骨的微通道设计,并与多目标优化算法耦合,以确定并提出基于观察到的热阻力和泵送功率的最佳溶液。提出了一种结合拉丁超立方体采样,基于机器学习的替代建模和多目标优化的工作流程。在寻找最佳替代物期间,考虑了随机森林,梯度增强算法和神经网络。我们证明了调整的神经网络可以做出准确的预测,并用于创建可接受的替代模型。与常规优化方法相比,优化解决方案在总体性能上显示出可忽略的差异。此外,解决方案是在原始时间的五分之一中计算的。在与对流微通道设计相同的压力极限下,生成的设计达到的温度低于10%以上。当受到温度的限制时,压降降低了25%以上。最后,通过采用Shapley添加说明技术研究了每个设计变量对热电阻和泵送功率的影响。总体而言,我们已经证明了所提出的框架具有优点,可以用作微通道散热器设计优化的可行方法。
translated by 谷歌翻译
血浆定义为物质的第四个状态,在高电场下可以在大气压下产生非热血浆。现在众所周知,血浆激活液体(PAL)的强和广谱抗菌作用。机器学习(ML)在医疗领域的可靠适用性也鼓励其在等离子体医学领域的应用。因此,在PALS上的ML应用可以提出一种新的观点,以更好地了解各种参数对其抗菌作用的影响。在本文中,通过使用先前获得的数据来定性预测PAL的体外抗菌活性,从而介绍了比较监督的ML模型。进行了文献搜索,并从33个相关文章中收集了数据。在所需的预处理步骤之后,将两种监督的ML方法(即分类和回归)应用于数据以获得微生物灭活(MI)预测。对于分类,MI分为四类,对于回归,MI被用作连续变量。为分类和回归模型进行了两种不同的可靠交叉验证策略,以评估所提出的方法。重复分层的K折交叉验证和K折交叉验证。我们还研究了不同特征对模型的影响。结果表明,高参数优化的随机森林分类器(ORFC)和随机森林回归者(ORFR)分别比其他模型进行了分类和回归的模型更好。最后,获得ORFC的最佳测试精度为82.68%,ORFR的R2为0.75。 ML技术可能有助于更好地理解在所需的抗菌作用中具有主要作用的血浆参数。此外,此类发现可能有助于将来的血浆剂量定义。
translated by 谷歌翻译
主动脉(COA)患者特异性计算流体动力学(CFD)研究的目的 - 在资源约束设置中的研究受到可用成像方式和速度数据采集的可用成像方式的限制。多普勒超声心动图被视为合适的速度获取方式,因为其可用性和安全性较高。这项研究旨在调查经典机器学习(ML)方法的应用,以创建一种适当且可靠的方法,用于从多普勒超声心动图图像中获得边界条件(BCS),用于使用CFD进行血液动力学建模。方法 - 我们提出的方法结合了ML和CFD,以模拟感兴趣区域内的血流动力学流动。该方法的关键特征是使用ML模型来校准CFD模型的入口和出口边界条件(BCS)。 ML模型的关键输入变量是患者心率,因为这是研究中测得的血管的时间变化的参数。在研究的CFD组件中使用ANSYS Fluent,而Scikit-Learn Python库则用于ML分量。结果 - 我们在干预前对严重COA的真实临床案例进行了验证。将我们的模拟的最大缩回速度与从研究中使用的几何形状获得的患者获得的测量最大骨质速度进行了比较。在用于获得BCS的5 mL模型中,顶部模型在测得的最大骨质速度的5 \%之内。结论 - 该框架表明,它能够考虑在测量之间考虑患者心率的变化。因此,当在每个血管上缩放心率时,可以在生理上逼真的BC计算,同时提供合理准确的溶液。
translated by 谷歌翻译
血流特征的预测对于了解血液动脉网络的行为至关重要,特别是在血管疾病(如狭窄)的存在下。计算流体动力学(CFD)提供了一种强大而有效的工具,可以确定包括网络内的压力和速度字段的这些特征。尽管该领域有许多研究,但CFD的极高计算成本导致研究人员开发新的平台,包括机器学习方法,而是以更低的成本提供更快的分析。在这项研究中,我们提出了一个深度神经网络框架,以预测冠状动脉网络中的流动行为,在存在像狭窄等异常存在下具有不同的性质。为此,使用合成数据训练人工神经网络(ANN)模型,使得它可以预测动脉网络内的压力和速度。培训神经网络所需的数据是从ABAQUS软件的特定特征的次数的CFD分析中获得了培训神经网络的数据。狭窄引起的血压下降,这是诊断心脏病诊断中最重要的因素之一,可以使用我们所提出的模型来了解冠状动脉的任何部分的几何和流动边界条件。使用Lad血管的三个实际几何形状来验证模型的效率。所提出的方法精确地预测了血流量的血流动力学行为。压力预测的平均精度为98.7%,平均速度幅度精度为93.2%。根据测试三个患者特定几何形状的模型的结果,模型可以被认为是有限元方法的替代方案以及其他难以实现的耗时数值模拟。
translated by 谷歌翻译
Building an accurate model of travel behaviour based on individuals' characteristics and built environment attributes is of importance for policy-making and transportation planning. Recent experiments with big data and Machine Learning (ML) algorithms toward a better travel behaviour analysis have mainly overlooked socially disadvantaged groups. Accordingly, in this study, we explore the travel behaviour responses of low-income individuals to transit investments in the Greater Toronto and Hamilton Area, Canada, using statistical and ML models. We first investigate how the model choice affects the prediction of transit use by the low-income group. This step includes comparing the predictive performance of traditional and ML algorithms and then evaluating a transit investment policy by contrasting the predicted activities and the spatial distribution of transit trips generated by vulnerable households after improving accessibility. We also empirically investigate the proposed transit investment by each algorithm and compare it with the city of Brampton's future transportation plan. While, unsurprisingly, the ML algorithms outperform classical models, there are still doubts about using them due to interpretability concerns. Hence, we adopt recent local and global model-agnostic interpretation tools to interpret how the model arrives at its predictions. Our findings reveal the great potential of ML algorithms for enhanced travel behaviour predictions for low-income strata without considerably sacrificing interpretability.
translated by 谷歌翻译
种植植被是降低沉积物转移率的实用解决方案之一。植被覆盖的增加可降低环境污染和沉积物的运输速率(STR)。由于沉积物和植被相互作用复杂,因此预测沉积物的运输速率具有挑战性。这项研究旨在使用新的和优化的数据处理方法(GMDH)的新版本(GMDH)预测植被覆盖的沉积物传输速率。此外,这项研究介绍了一种用于预测沉积物传输速率的新集合模型。模型输入包括波高,波速,密度覆盖,波力,D50,植被盖的高度和盖茎直径。独立的GMDH模型和优化的GMDH模型,包括GMDH Honey Badger算法(HBA)GMDH大鼠群群算法(RSOA)VGMDH正弦余弦算法(SCA)和GMDH颗粒swarm swarm优化率(GMDH-PSO),用于预测沉积率(GMDH-PSO) 。作为下一步,使用独立的GMDH的输出来构建集合模型。合奏模型的MAE为0.145 m3/s,而GMDH-HBA,GMDH-RSOA,GMDH-SCA,GMDH-PSOA和GMDH的MAE在测试水平为0.176 M3/s,0.312 M3/s,0.367/s,0.367 M3/s,0.498 m3/s和0.612 m3/s。集合模型的Nash Sutcliffe系数(NSE),GMDH-HBA,GMDH-RSOA,GMDH-SCA,GMDH-PSOA和GHMDH分别为0.95 0.93、0.89、0.89、0.86、0.86、0.82和0.76。此外,这项研究表明,植被覆盖的沉积物运输速率降低了90%。结果表明,合奏和GMDH-HBA模型可以准确预测沉积物的传输速率。根据这项研究的结果,可以使用IMM和GMDH-HBA监测沉积物的传输速率。这些结果对于管理和规划大盆地的水资源很有用。
translated by 谷歌翻译
天然气管道中的泄漏检测是石油和天然气行业的一个重要且持续的问题。这尤其重要,因为管道是运输天然气的最常见方法。这项研究旨在研究数据驱动的智能模型使用基本操作参数检测天然气管道的小泄漏的能力,然后使用现有的性能指标比较智能模型。该项目应用观察者设计技术,使用回归分类层次模型来检测天然气管道中的泄漏,其中智能模型充当回归器,并且修改后的逻辑回归模型充当分类器。该项目使用四个星期的管道数据流研究了五个智能模型(梯度提升,决策树,随机森林,支持向量机和人工神经网络)。结果表明,虽然支持向量机和人工神经网络比其他网络更好,但由于其内部复杂性和所使用的数据量,它们并未提供最佳的泄漏检测结果。随机森林和决策树模型是最敏感的,因为它们可以在大约2小时内检测到标称流量的0.1%的泄漏。所有智能模型在测试阶段中具有高可靠性,错误警报率为零。将所有智能模型泄漏检测的平均时间与文献中的实时短暂模型进行了比较。结果表明,智能模型在泄漏检测问题中的表现相对较好。该结果表明,可以与实时瞬态模型一起使用智能模型,以显着改善泄漏检测结果。
translated by 谷歌翻译
The hydrodynamic performance of a sea-going ship varies over its lifespan due to factors like marine fouling and the condition of the anti-fouling paint system. In order to accurately estimate the power demand and fuel consumption for a planned voyage, it is important to assess the hydrodynamic performance of the ship. The current work uses machine-learning (ML) methods to estimate the hydrodynamic performance of a ship using the onboard recorded in-service data. Three ML methods, NL-PCR, NL-PLSR and probabilistic ANN, are calibrated using the data from two sister ships. The calibrated models are used to extract the varying trend in ship's hydrodynamic performance over time and predict the change in performance through several propeller and hull cleaning events. The predicted change in performance is compared with the corresponding values estimated using the fouling friction coefficient ($\Delta C_F$). The ML methods are found to be performing well while modelling the hydrodynamic state variables of the ships with probabilistic ANN model performing the best, but the results from NL-PCR and NL-PLSR are not far behind, indicating that it may be possible to use simple methods to solve such problems with the help of domain knowledge.
translated by 谷歌翻译
Surrogate models are necessary to optimize meaningful quantities in physical dynamics as their recursive numerical resolutions are often prohibitively expensive. It is mainly the case for fluid dynamics and the resolution of Navier-Stokes equations. However, despite the fast-growing field of data-driven models for physical systems, reference datasets representing real-world phenomena are lacking. In this work, we develop AirfRANS, a dataset for studying the two-dimensional incompressible steady-state Reynolds-Averaged Navier-Stokes equations over airfoils at a subsonic regime and for different angles of attacks. We also introduce metrics on the stress forces at the surface of geometries and visualization of boundary layers to assess the capabilities of models to accurately predict the meaningful information of the problem. Finally, we propose deep learning baselines on four machine learning tasks to study AirfRANS under different constraints for generalization considerations: big and scarce data regime, Reynolds number, and angle of attack extrapolation.
translated by 谷歌翻译
This work presents a set of neural network (NN) models specifically designed for accurate and efficient fluid dynamics forecasting. In this work, we show how neural networks training can be improved by reducing data complexity through a modal decomposition technique called higher order dynamic mode decomposition (HODMD), which identifies the main structures inside flow dynamics and reconstructs the original flow using only these main structures. This reconstruction has the same number of samples and spatial dimension as the original flow, but with a less complex dynamics and preserving its main features. We also show the low computational cost required by the proposed NN models, both in their training and inference phases. The core idea of this work is to test the limits of applicability of deep learning models to data forecasting in complex fluid dynamics problems. Generalization capabilities of the models are demonstrated by using the same neural network architectures to forecast the future dynamics of four different multi-phase flows. Data sets used to train and test these deep learning models come from Direct Numerical Simulations (DNS) of these flows.
translated by 谷歌翻译
激光加工是一种高度灵活的非接触式制造技术,在学术界和行业中广泛使用。由于光和物质之间的非线性相互作用,模拟方法非常重要,因为它们通过理解激光处理参数之间的相互关系来帮助增强加工质量。另一方面,实验处理参数优化建议对可用处理参数空间进行系统且耗时的研究。一种智能策略是采用机器学习(ML)技术来捕获Picsecond激光加工参数之间的关系,以找到适当的参数组合,以创建对工业级氧化铝陶瓷的所需削减,并具有深层,平滑和无缺陷的模式。激光参数,例如梁振幅和频率,扫描仪的传递速度以及扫描仪与样品表面的垂直距离的速度,用于预测深度,最高宽度和底部宽度使用ML型号雕刻通道。由于激光参数之间的复杂相关性,因此表明神经网络(NN)是预测输出最有效的。配备了ML模型,该模型可以捕获激光参数与雕刻通道尺寸之间的互连,可以预测所需的输入参数以实现目标通道几何形状。该策略大大降低了开发阶段实验激光加工的成本和精力,而不会损害准确性或性能。开发的技术可以应用于各种陶瓷激光加工过程。
translated by 谷歌翻译
预测经济的短期动态 - 对经济代理商决策过程的重要意见 - 经常在线性模型中使用滞后指标。这通常在正常时期就足够了,但在危机期间可能不足。本文旨在证明,在非线性机器学习方法的帮助下,非传统和及时的数据(例如零售和批发付款)可以为决策者提供复杂的模型,以准确地估算几乎实时的关键宏观经济指标。此外,我们提供了一组计量经济学工具,以减轻机器学习模型中的过度拟合和解释性挑战,以提高其政策使用的有效性。我们的模型具有付款数据,非线性方法和量身定制的交叉验证方法,有助于提高宏观经济的启示准确性高达40 \% - 在COVID-19期间的增长较高。我们观察到,付款数据对经济预测的贡献很小,在低和正常增长期间是线性的。但是,在强年或正增长期间,付款数据的贡献很大,不对称和非线性。
translated by 谷歌翻译
台湾对全球碎片流的敏感性和死亡人数最高。台湾现有的碎屑流警告系统,该系统使用降雨量的时间加权度量,当该措施超过预定义的阈值时,会导致警报。但是,该系统会产生许多错误的警报,并错过了实际碎屑流的很大一部分。为了改善该系统,我们实施了五个机器学习模型,以输入历史降雨数据并预测是否会在选定的时间内发生碎屑流。我们发现,随机的森林模型在五个模型中表现最好,并优于台湾现有系统。此外,我们确定了与碎屑流的发生密切相关的降雨轨迹,并探索了缺失碎屑流的风险与频繁的虚假警报之间的权衡。这些结果表明,仅在小时降雨数据中训练的机器学习模型的潜力可以挽救生命,同时减少虚假警报。
translated by 谷歌翻译
Different machine learning (ML) models are trained on SCADA and meteorological data collected at an onshore wind farm and then assessed in terms of fidelity and accuracy for predictions of wind speed, turbulence intensity, and power capture at the turbine and wind farm levels for different wind and atmospheric conditions. ML methods for data quality control and pre-processing are applied to the data set under investigation and found to outperform standard statistical methods. A hybrid model, comprised of a linear interpolation model, Gaussian process, deep neural network (DNN), and support vector machine, paired with a DNN filter, is found to achieve high accuracy for modeling wind turbine power capture. Modifications of the incoming freestream wind speed and turbulence intensity, $TI$, due to the evolution of the wind field over the wind farm and effects associated with operating turbines are also captured using DNN models. Thus, turbine-level modeling is achieved using models for predicting power capture while farm-level modeling is achieved by combining models predicting wind speed and $TI$ at each turbine location from freestream conditions with models predicting power capture. Combining these models provides results consistent with expected power capture performance and holds promise for future endeavors in wind farm modeling and diagnostics. Though training ML models is computationally expensive, using the trained models to simulate the entire wind farm takes only a few seconds on a typical modern laptop computer, and the total computational cost is still lower than other available mid-fidelity simulation approaches.
translated by 谷歌翻译
机器学习(ML)是指根据大量数据预测有意义的输出或对复杂系统进行分类的计算机算法。 ML应用于各个领域,包括自然科学,工程,太空探索甚至游戏开发。本文的重点是在化学和生物海洋学领域使用机器学习。在预测全球固定氮水平,部分二氧化碳压力和其他化学特性时,ML的应用是一种有前途的工具。机器学习还用于生物海洋学领域,可从各种图像(即显微镜,流车和视频记录器),光谱仪和其他信号处理技术中检测浮游形式。此外,ML使用其声学成功地对哺乳动物进行了分类,在特定的环境中检测到濒临灭绝的哺乳动物和鱼类。最重要的是,使用环境数据,ML被证明是预测缺氧条件和有害藻华事件的有效方法,这是对环境监测的重要测量。此外,机器学习被用来为各种物种构建许多对其他研究人员有用的数据库,而创建新算法将帮助海洋研究界更好地理解海洋的化学和生物学。
translated by 谷歌翻译
Machine learning-based modeling of physical systems has experienced increased interest in recent years. Despite some impressive progress, there is still a lack of benchmarks for Scientific ML that are easy to use but still challenging and representative of a wide range of problems. We introduce PDEBench, a benchmark suite of time-dependent simulation tasks based on Partial Differential Equations (PDEs). PDEBench comprises both code and data to benchmark the performance of novel machine learning models against both classical numerical simulations and machine learning baselines. Our proposed set of benchmark problems contribute the following unique features: (1) A much wider range of PDEs compared to existing benchmarks, ranging from relatively common examples to more realistic and difficult problems; (2) much larger ready-to-use datasets compared to prior work, comprising multiple simulation runs across a larger number of initial and boundary conditions and PDE parameters; (3) more extensible source codes with user-friendly APIs for data generation and baseline results with popular machine learning models (FNO, U-Net, PINN, Gradient-Based Inverse Method). PDEBench allows researchers to extend the benchmark freely for their own purposes using a standardized API and to compare the performance of new models to existing baseline methods. We also propose new evaluation metrics with the aim to provide a more holistic understanding of learning methods in the context of Scientific ML. With those metrics we identify tasks which are challenging for recent ML methods and propose these tasks as future challenges for the community. The code is available at https://github.com/pdebench/PDEBench.
translated by 谷歌翻译
有希望的方法来改善气候模型中的云参数化,因此气候预测是使用深度学习与来自Storm-解析模型(SRM)模拟的培训数据结合使用。 ICOSAHEDRAL非静水压(图标)建模框架允许模拟从数值天气预报到气候投影,使其成为开发基于神经网络(NN)的子网比例过程的参数化的理想目标。在图标框架内,我们通过基于逼真的区域和全局图标SRM模拟培训基于NN的云覆盖参数化。我们设置了三种不同类型的NNS,其垂直局部程度不同,它们假设从粗粒粒度大气状态变量诊断云盖。 NNS精确地从粗粒数据中估计子网格尺度云覆盖,该数据具有与其训练数据相似的地理特征。此外,全球培训的NNS可以再现区域SRM仿真的子网格级云覆盖。使用基于游戏理论的可解释性库福芙添加剂解释,我们识别特定湿度和云冰上的过分传播,以及我们基于列的NN不能从全局到区域粗粒度SRM数据完全概括的原因。该解释工具还有助于可视化区域和全球训练的基于列的NNS之间的特征重要性的相似性和差异,并在其云覆盖预测和热力学环境之间揭示了本地关系。我们的结果表明,深度学习的潜力从全球SRMS获得准确但可解释的云覆盖参数化,并表明基于邻域的模型可能是精度和概括性之间的良好折衷。
translated by 谷歌翻译
自从Navier Stokes方程的推导以来,已经有可能在数值上解决现实世界的粘性流问题(计算流体动力学(CFD))。然而,尽管中央处理单元(CPU)的性能取得了迅速的进步,但模拟瞬态流量的计算成本非常小,时间/网格量表物理学仍然是不现实的。近年来,机器学习(ML)技术在整个行业中都受到了极大的关注,这一大浪潮已经传播了流体动力学界的各种兴趣。最近的ML CFD研究表明,随着数据驱动方法的训练时间和预测时间之间的间隔增加,完全抑制了误差的增加是不现实的。应用ML的实用CFD加速方法的开发是剩余的问题。因此,这项研究的目标是根据物理信息传递学习制定现实的ML策略,并使用不稳定的CFD数据集验证了该策略的准确性和加速性能。该策略可以在监视跨耦合计算框架中管理方程的残差时确定转移学习的时间。因此,我们的假设是可行的,即连续流体流动时间序列的预测是可行的,因为中间CFD模拟定期不仅减少了增加残差,还可以更新网络参数。值得注意的是,具有基于网格的网络模型的交叉耦合策略不会损害计算加速度的仿真精度。在层流逆流CFD数据集条件下,该模拟加速了1.8次,包括参数更新时间。此可行性研究使用了开源CFD软件OpenFOAM和开源ML软件TensorFlow。
translated by 谷歌翻译
在热分析和低温热交换器的几何设计过程中,强迫沸腾现象的准确降低估计很重要。但是,当前预测压降的方法存在两个问题之一:缺乏对不同情况的准确性或概括。在这项工作中,我们介绍了相关的信息神经网络(COINN),这是应用人工神经网络(ANN)技术的新范式,结合了成功的压降相关性,作为预测微质混合压力下降的绘制工具 - 通道。所提出的方法是受转移学习的启发,该方法在减少数据集的深度学习问题中高度使用。我们的方法通过将Sun&Mishima相关性的知识传递给ANN来提高ANN的性能。具有物理和现象学对微通道压力下降的相关性大大提高了ANN的性能和概括能力。最终结构由三个输入组成:混合蒸气质量,微通道内径和可用的压降相关性。结果表明,使用相关的信息方法获得的好处预测用于训练的实验数据和后验测试,平均相对误差(MRE)为6%,低于Sun&Mishima相关性13%。此外,这种方法可以扩展到其他混合物和实验设置,这是使用ANN用于传热应用的其他方法中的缺少特征。
translated by 谷歌翻译