1. Introduction
Additive manufacturing has been considered an important and growing form of manufacturing for many years. The standard ISO/ASTM 52900:2021 [
1] defines it as the process by which, starting from the data of a 3D model, a part is manufactured layer by layer. However, the way to comply with this definition is extremely varied and depends on different technologies. This standard divides these technologies into seven categories, but only four of them can directly use their raw material in powder form: binder jetting, directed energy deposition, material extrusion, and powder bed fusion.
The materials used in these technologies can be metallic, polymeric, or ceramic, but all have in common their high quality and consequent cost.
Taking directed energy deposition as an example, this technology uses a combination of inner gas (normally argon) as a clean atmosphere to avoid oxidation, feeding material (metallic or polymeric powder [
2]), and a power source to melt the powder in a controlled manner (a laser or electron beam). A schematic of a DED powder system is shown in
Figure 1.
This technology divides the powder feed into three moments [
3]: metering, conveyance, and delivery. Each moment has different possible methods (mechanical, pneumatic, vibration-assisted, or gravitational methods, for example), all of them with different control necessities. However, a high accurate spray control is needed because part of the sprayed powder is not used [
4]. This manufacturing system could have multiple nozzles, being able to mix different materials to obtain Functionally Graded Materials (FGMs) [
5,
6].
Analyzing the three moments in the powder feed, a novel solution could be the use of Venturi injectors, which allow the powder to be sucked from one point and pushed to where it is needed in a controlled manner, unifying the metering and conveyance systems.
From analyzing the solutions on the market that use these technologies, it can be seen that the raw materials in powder format tend to have very small particle sizes and are available in all densities. These characteristics can greatly influence the design of Venturi injectors and must therefore be taken into account when sizing them.
The most studied point in the literature is the net efficiency of the Venturi injector, its mean, and the lowest possible pressure drops. In this direction, Zerpa et al. [
7] studied the geometry effects of the divergent and convergent sections in classical injectors, as did also Almeida [
8], but this one with the aim of improving the propulsion of the fluid at the injector outlet. Zhang [
9] focused on studying how the dimensional parameters affect the pressure distribution inside the injector, concluding that the throat length does not affect the pressure distribution inside the injector. Also analyzing these variables, in [
10], aspect ratios were established that sought the maximum allowable differential pressure for the injector to operate at the maximum flow rate of the injector. However, Park [
11] established a maximum differential pressure criterion as a function of Reynolds number.
A typical effect sought in the use of Venturi injectors is the generation of vacuum, but at a certain point the depression can produce cavitation, and given the importance of this effect, Shi et al. and Bermejo et al. [
12,
13] studied how this cavitation affects and its possible uses. One of these uses is fluid mixing; thus, Shi and Nikrityuk [
14,
15] studied how this cavitation affects the mixing process.
O’Hern et al. [
16] established a method for the design of these injectors, while Sobenko et al. [
17] proposed a model for predicting the performance of a Venturi injector, in addition to analyzing the flow pressure ratios that generated the highest level of instability in the Venturi injector.
On the other hand, there are studies focused on powder transport or injection. This use of Venturi injectors has been well established for many years [
18], but progress is still being made in understanding the internal phenomena of the injector. When powder is introduced both in virtual tests [
19] as well as in real tests [
20], it should be noted that the requirements for fluid transport are different from those for powder transport, and so the design of injector geometry continues to be studied and advanced, as did by Xu et al. [
21]. Of course, it is necessary to study how the sprayed powder behaves depending on its use [
22,
23].
In the case of powder transport, it must also be taken into account how particle morphology affects injector performance [
24].
In terms of injector optimization, Expósito et al. [
25] established a procedure applied to classical injectors focused on fluid mixing by using genetic algorithms, and Wang et al. [
26] performed another optimization, also applying it to fluid mixing, but using a somewhat different Venturi injector design and machine learning algorithms as the optimization method.
In all the above studies, no work can be found that focuses on the design of an optimal Venturi with a view to maximizing powder transport as a priority. It is also noted that the powder transport studies focus on Venturi injectors that feed powder by gravity, not by suction.
In addition, if the DED manufacturing process requires an inert gas (normally at high pressure), the powder feed needs to be at lower pressure to guarantee the fast cleaning of the atmosphere. Thus, the transport of powder at atmospheric pressure is a good option, but no studies have been found that focus on the design of low-pressure Venturi injectors (less than 100 kPa) [
18].
Based on these detected shortcomings, this study optimized the different dimensional parameters of a Venturi injector to maximize the amount of powder capable of suction and transport for a specific flow in a range of low pressures and with high transport precision for the development of a mixing powder device for Functionally Graded Additive Manufacturing.
2. Materials and Methods
This work presents a methodology for optimizing the dimensions of Venturi injectors with low pressure for the transportation of powder with high accuracy. To achieve this, this section presents the description of the problem and the main dimensions of Venturi injectors; then, the modeling of the problem (both from the point of view of the flow, CFD, particle interaction, DEM, and coupling between them); and finally, the optimization tools used to obtain the optimal design. Alongside this, the final section explains the statistical tools used to determine the significance of variables with the results obtained. This process follows the flowchart shown in
Figure 2.
2.1. Problem Description
The objective of this work is to establish a sizing methodology for optimized Venturi injectors, using particle and flow characteristics as a reference.
The mean particle diameter of the powder to be conveyed and the flow velocity were used as reference constants.
In this case, a powder named Powder-ID256-Scale_1.00 was used, which can be found in the database of representative but not real calibrated powders in Altair Edem software
® (2022 version) and has the characteristics presented in
Table 1. For the flow rate, a volumetric flow rate of 0.01
was considered.
Since it is not entirely clear which factors influence powder suction, the dimensional variables shown in
Table 2 were taken as the objects of our study. These variables are depicted in
Figure 3.
Due to the physical constraints of possible laboratory manufacture, although the variables D1, d1, and D2 have a percentage relationship with other variables, if the result of this relation is greater than 34 mm, this value will be taken as the dimension.
2.2. Modeling the Problem
To carry out this study, a simulated environment was developed. For these simulations, 3 interactions must be taken into account: the fluid, the particle–particle interaction, and the interaction between the fluid and the particles. In addition, the simulation interval time must be the minimum possible with the objective of reducing the computational time, but enough to guarantee powder ejection from the Venturi. In this study, that time was 2 s.
2.2.1. Modeling the Fluid Motion
For fluid modeling, the fluid was taken as a continuous element and was governed by the Navier–Stokes continuity equations. In general, applying the conservation of mass to a control volume, the equation in differential form corresponds to Equation (1):
where
ρ is the density of the fluid;
is the time; and
is the flow velocity vector.
In turn, the conservation of motion quantity equation is obtained by applying Newton’s second law to the same control volume, where Equation (2) governs it, and its components can be viscous, pressure, gravity, centrifugal forces, Coriolis forces, etc.:
where
is the pressure;
is the gravity; and
is the viscous stress vector corresponding to Equation (3), where the vector
is decomposed in the
,
, and
directions with the following subscript:
where
is the dynamic viscosity of the fluid.
In addition, the Spalart–Allmaras turbulent flow model [
27] was added, which is governed by Equation (4):
where
and
. Also,
and
are the production and destruction terms of the modified turbulent viscosity, corresponding to Equations (5) and (9), respectively:
with
where
is the rotation tensor;
is the distance from the nearest wall;
is the Von Karman’s constant (0.41); and
.
with
Moreover, for the modeling of the turbulent viscosity, Equation (13) was used:
with
This model was completed with the coefficients , , , and .
2.2.2. Modeling of Particle Interaction
Different mathematical models can be used for particle modeling, depending on the particle characteristics.
When talking about polymeric particles, it is assumed that they have a certain level of elasticity. This means that when the particles interact with the walls, they undergo some form of deformation; in the case of particle-to-particle interactions, this deformation causes the distance between their centers to be reduced due to what is considered an overlap between the two particles
Figure 4b.
In
Figure 4a, these interactions are represented at the point of contact of two particles by a parallel connection between a spring and a damper.
Depending on the mathematical model used, the energy transferred by these interactions is represented to a greater or lesser extent in the equations of motion. An example of this is shown in
Figure 5, where the contact force–motion functions are depicted:
Figure 5a is the representation of the most basic contact model (Hert–Mindlin contact model) [
28,
29,
30,
31,
32], and
Figure 5b depicts the Edinburgh Elasto-Plastic Adhesion contact model.
The equations governing the Hert–Mindlin (no slip) model are summarized in
Table 3 [
33].
Although each of these models is more complete than the previous one, the complexity of these models results in a high computational cost, so it is not always more effective to choose the most complete model.
In the case of the present work, the Hert–Mindlin (no-slip) model was used for the particle–wall interactions, and the Edinburgh Elasto-Plastic Adhesion model was employed for the interaction between particles.
When two particles come into contact, the force is distributed into two components called the normal force (
) and tangential force (
), as can be seen in
Figure 4b. The definition of each of these forces depends on the model.
In the case of the Edinburgh Elasto-Plastic Adhesion model, some factors are added to account for particle plasticity. The governing equations are summarized in
Table 4 [
34].
2.2.3. The Particle Flow Coupling Model
To study the behavior of the particles in the fluid in motion, a joint computational fluid dynamics (CFD) and discrete element method (DEM) simulation was carried out, using the models explained above, in each simulation time step. For this purpose, the software tools Altair Edem
® (2022 version) and Altair Hyperworks CFD
® (2022 version) were run together, following the workflow shown in
Figure 6.
In this way, the workflow starts in the Hyperworks CFD tool, which calculates the volume fractions, forces, and moments of the fluid; it then updates the information for that time step and transfers it to the EDEM tool.
In the EDEM tool, the forces and contacts of the particles are calculated; the position and velocity of the particles are updated; and this information is transferred to the Hyperworks CFD tool.
This cycle was repeated with each time step until the end of the set test time.
To obtain an accurate calculation of the trajectory of these particles, it is necessary to use a mathematical entrainment model that transfers the information between the two phases of the test (fluid and powder). In this study, a model [
35,
36] that combines the Ergun [
37] and Wen-Yu [
38] models was used to evaluate the tensile force,
according to Equation (16):
where
where
is the gas velocity vector;
is the volume of particle I;
corresponds to the solid velocity vector; and
is the free volume fraction.
2.3. Optimization Using Genetic Algorithms
A process for choosing what variable is modified and the measure of this change is needed. In this study, the process selected is the use of genetic algorithms.
Genetic algorithms are a mathematical assimilation of the law of natural selection in nature [
39,
40]. In this case, 100 generations with 100 individuals each were applied. A tournament selection was applied by selecting two random individuals and comparing the fitness function value (using the Kriging metamodel interpolation method to estimate the possible responses [
41,
42]) to store the best one in the intermediate population (100 tournaments until obtaining an intermediate population of 100 individuals). Half of the intermediate population (50% crossover rate) was randomly selected and combined to obtain offspring (arithmetic crossover), while the other half of the intermediate population was maintained. Additionally, a 60% mutation rate was applied, as well as elitism (the worst individual of the new populations was always replaced by the best one of the previous populations). To create these metamodels, the results previously obtained in a design of experiments were used as a reference.
The Kriging metamodel uses an exponential correlation model and a polynomial regression model. However, polynomial regression is able to change the order from 0 to 2, depending on the recorded data. When you have a lower-order polynomial, it requires fewer reference values and is easier to calculate, but it has less precision. For this, the algorithm starts by trying to make use of a second-order regression model to predict each response; if this fails, it will automatically use a first-order regression model, and if this fails, it will go back down using a zero-order regression model. This automatic order change ensures that you are always using the best possible regression model with the data you have.
The modified Latin hypercube algorithm is used to generate a set of initial cases for initial interpolations [
39]. First, a combination is generated with all the variables at the minimum of their range, another combination with all variables at the midpoint of their range, and another combination with all the variables at their maximum point, which implies 3 designs. Subsequently, the Latin hypercube algorithm is applied to add “n” points, where “n” is the number of design variables. Therefore, this algorithm divides the range of each variable into equal parts, selects a random value for each variable, and eliminates the corresponding range for each of the variables from the list of possible values to be used in the next combination of variables. This process is repeated until “n” points are added. This means that the optimization will start with a total of 3+n combinations, which in this study means a total of 12 initial combinations.
Since the Kriging metamodel interpolation method gives better results interpolating than extrapolating, the Latin hypercube algorithm was modified so that all generated values corresponding to the first or last rank of each variable are modified and set to the minimum or maximum value, respectively. In this way, these data are shifted to the contour of the search space, promoting data interpolation over extrapolation.
In this study, the Venturi design variables explained in
Section 2.1 were used as the variables to be optimized, while the mass of powder discharged in two seconds (Obj
1) and the pressure drop produced between the inlet and the outlet of the Venturi injector (Obj
2) are the response variables to maximize and minimize, respectively. To achieve this in a mono-objective approach, the objective function was chosen to transform the scale of the powder discharge and pressure drop results by normalizing them. For this purpose, Equation (20) transforms the minimum value of each variable into 0 and the maximum value into 1, scaling the rest of the results. This transformation is called normalization.
Thus, the objective function will be the maximization of Equation (21):
The mean absolute percentage error (also known as MAPE) was used as a stopping criterion in each iteration [
41], which measures, in percentage terms, the absolute error committed by the response variables as a whole. This indicator responds to Equation (22):
where
is the number of response variables to optimize;
is the reference value (those obtained in the simulations to check); and
is the value to be compared (those given by the algorithm as an estimated value according to the data available for training the Kriging metamodel).
When this criterion is below 4%, the optimization will be considered good since the subsequent numerical simulation of the solution proposed by the genetic algorithm approximately matches (MAPE < 4%) the estimated results of the algorithm. However, if the iteration reaches 4% but better results have been obtained in previous iterations, it continues iterating until the objective function is the best of all cases studied while the maximum MAPE criterion allowed is met. Sometimes the stopping criterion is not reached after many iterations. In these cases, every 10 iterations without reaching the stopping criterion, the results of the MAPE are analyzed. If the trend of these results is towards 4%, 10 more iterations are performed. In case the results start to fluctuate with no improving trend, the best result of the objective function in the last 10 iterations will be assumed as the optimum.
The duty cycle for each iteration can be seen in
Figure 7.
2.4. Determination of Significance of Variables
In order to study the relative importance, from a statistical point of view, of the different design variables considered, a multiple linear regression analysis with standardized coefficients was carried out to model the behavior of powder discharge.
2.4.1. Multiple Linear Regression
The objective of regression analysis is to mathematically model the behavior of a response variable () as a function of one or more independent or predictor variables ().
The multiple linear regression model can be written in matrix notation as follows:
where
The vector contains the coefficients of the model, where is the ordinate at the origin and , is the average effect that a one-unit increase in the predictor variable has on the dependent variable , when all the other regressor variables are held fixed or constant. The vector contains the errors or residuals, where is the difference between the observed value and the value estimated by the model. In addition, .
Since the magnitude of each regression coefficient depends on the units in which the predictor variable to which it corresponds is measured, in order to determine the impact of each variable on the model, standardized coefficients were used, which were obtained by standardizing (subtracting the mean and dividing by the standard deviation) the predictor variables prior to the model’s fitting.
Multiple linear correlation models require the following conditions:
Linearity between the independent variables of the model and the response variable.
Normality of the residuals: It is assumed that the residuals are normally distributed with a zero mean. This is usually checked by means of graphical methods (histograms, box-and-whisker plots, or quantile–quantile (Q–Q) plots) as well as normality hypothesis tests such as, for example, the Shapiro–Wilk test, which is applicable when analyzing small samples (composed of less than 50 elements) or the Anderson–Darling test, which is a non-parametric test on sample data (with more than 7 elements) coming from a specific distribution.
Homogeneity of the variance of the residuals (homoscedasticity): To check this, the residuals are plotted. If the variance is constant, they are randomly distributed with the same dispersion and without any specific pattern. One can also resort to homoscedasticity tests such as the Breusch–Pagan test, which only detects linear forms of heteroscedasticity, or the Goldfeld–Quandt test, which compares the variances of two sub-models separated by a specified break point and rejects if the variances differ.
The residuals are independent of each other: The Durbin–Watson test allows for the diagnosis of the presence of a correlation between consecutive residuals ordered in time, which is a possible manifestation of a lack of independence.
There are no outliers with a high influence. That is, the regression model is not strongly influenced by one or more outlier data points because this would raise doubts about the adequacy of the model and the reliability of the data in some cases. Cook’s distance allows for detecting outliers with a high influence. For a sample of elements and a model of independent variables, a Cook’s distance greater than the median of an F-distribution with and degrees of freedom, with , is considered of concern.
Uncorrelated predictors: In multiple linear models, the predictors must be independent, meaning there must be no collinearity between them. Tolerance and the variance inflation factor () are two parameters that quantify the same thing (one is the inverse of the other). The reference limits that are usually used are: : absence of collinearity; : regression may be affected by some collinearity; : cause for concern; and : serious collinearity.
Parsimony: This term refers to the fact that the best model is the one that can most accurately explain the variability observed in the response variable using the least number of predictors.
Good models are those that meet the most criteria for goodness-of-fit. There will always be circumstances where failure to meet any of the criteria will not necessarily make the model unfeasible from a practical point of view [
42].
2.4.2. Choice of Predictors to Generate the Best Model
When selecting the predictors to be part of the model, several methods can be followed, among them the so-called stepwise methods, which consist of iteratively adding and/or removing predictors in the regression model in order to find the subset of variables in the dataset that results in a model that reduces the prediction error. There are three stepwise regression strategies:
Forward selection, which starts with no predictors in the model, iteratively adds the most contributing predictors, and stops when the improvement is no longer statistically significant.
Backward selection (or backward elimination), which starts with all predictors in the model (full model), iteratively eliminates the least contributing predictors, and stops when you have a model in which all predictors are statistically significant.
Stepwise selection (or sequential substitution), which is a combination of the forward and backward selections. You start with no predictors and sequentially add the most contributing predictors (as in forward selection). After each new variable is added, any variable that no longer provides an improvement in model fit (as in backward selection) is removed.
The step-by-step method requires some mathematical criteria to determine whether the model gets better or worse with each addition or removal. There are several parameters or metrics that can be used, the most important of which are:
, Mallows’
, the Akaike Information Criterion (AIC), and the Bayesian Information Criterion (BIC) [
43].
When there are terms in the model that do not contribute significantly to the model, tends to be smaller than . Therefore, it is desirable to refine the model. In general, to speak of a model that has a satisfactory fit, it is necessary that the coefficients and have values greater than 0.7. However, when comparing different models for the same dataset, a lower AIC/BIC score is better.
2.4.3. The Lack-of-Fit Test
The F-test is useful for checking that the model as a whole performs better than chance. If a regression model does not produce a significant F-test result, then you probably do not have a very good regression model (or, quite possibly, you do not have very good data). However, while failing this test is a fairly strong indicator that the model has problems, passing the test (i.e., rejecting the null hypothesis) does not imply that the model is good.
Most computer programs specialized in statistics include procedures to perform both simple and multiple regression analyses and usually include variable selection techniques. In this work, we used the free software Jamovi [
44] based on the statistical language R [
45,
46].
4. Conclusions
After evaluating the different combinations of the dimensional variables required for the manufacture of low-velocity Venturi injectors using combined CFD and DEM simulations, a genetic algorithm was used to achieve an optimal combination of variables for sizing a Venturi injector based on mean particle diameter and flow rate.
From analyzing these results, it can be seen that, by aiming for the lowest possible pressure drop, the Venturi convergence and divergence angles (α and β) are within the ranges of the standard for the manufacture of classic Venturi injectors. Therefore, they can be considered valid.
Furthermore, from the statistical study, it can be concluded that the variables that most significantly affect powder suction are D3, d1, and d2 (in that order), which makes it clear that d1 and d2 are important because they are the two variables on whose size the vacuum generated in the mixing chamber depends, and at equal pressure, the larger the suction diameter (D3), the more physical space there is for the particles to rise.
This suggests that the switching variables D1 and D2 can act independently, allowing them to be adapted to the system where the injector is to be installed.
It is also confirmed that throat length does not affect powder deposition [
9], thus making it possible to reduce the size of the injector. Additionally, the position of the nozzle inside the mixing chamber is similar to that demonstrated by Xu et al. [
21].
On the other hand, the simulations of the optimal Venturi configuration with different inlet flow demonstrated a good linearity between inlet flow and powder discharge, which means that it is possible to control the mass flow and, consequently, the powder discharge. With this approach, different Venturi injectors could be integrated for the development of accurate mixing powder devices intended for Functionally Graded Additive Manufacturing. However, this linearity was only demonstrated for flow rates lower than the one used for optimization (0.01 m
3⁄s), so the optimal parameters should be recalculated for higher flow rates. Since the objective of this study is the use of this optimized injector for a DED system, the recommendation is to optimize the injector for the maximum flow rate required in each case. In this optimization process, we found better ranges of pressure drop than observed by Zerpa et al. [
7], and better linearity than observed by Almeida et al. [
8] about the press drop–flow relationship. We also demonstrated that, in the case of powder transportation, the relation between the nozzle and throat diameters is above the found in the bibliography. Comparing our results with those obtained by Wang et al. [
26] using machine learning, we can see how, while the input and outlet angles are similar, the proportional dimensions of the nozzle and the throat are different, confirming that the optimization is not the same when the feed process is by gravity or suction.
It should be noted that the mathematical models used in this work are fully validated according to the literature. However, future research will be carried out to experimentally validate these results. On the other hand, this optimization procedure is based on the particle diameter, so, in future research, it remains to be seen its effectiveness when applied to particles with different properties such as density, cohesivity, or diameters, even opening the possibility to study a range of optimal dimensions for more than one powder in the same injector.
Having demonstrated that it is possible to use Venturi injectors to unify the metering and conveyance in DEM systems, it has the following advantages over traditional systems: being a unified system whose implementation should be easier and cheaper; easy installation of more nozzles, thus increasing the possibilities of Functionally Graded Materials; and easy change of materials (cleaning and feeding). In view of the results obtained, the use of Venturi injectors could be studied in other technologies such as material extrusion or powder bed fusion, changing the feeding method, and giving the possibility to work with Functionally Graded Materials in different ways.