Previous Article in Journal
Does Audit Oversight Quality Reduce Insolvency Risk, Systematic Risk, and ROA Volatility? The Role of Institutional Ownership
Previous Article in Special Issue
On Smoothing and Habit Formation of Variable Life Annuity Benefits
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Extended Least Squares Making Evident Nonlinear Relationships between Variables: Portfolios of Financial Assets

by
Pierpaolo Angelini
Dipartimento di Scienze Statistiche, Università La Sapienza, 00185 Roma, Italy
J. Risk Financial Manag. 2024, 17(8), 336; https://doi.org/10.3390/jrfm17080336 (registering DOI)
Submission received: 22 June 2024 / Revised: 26 July 2024 / Accepted: 31 July 2024 / Published: 2 August 2024

Abstract

:
This research work extends the least squares criterion. The regression models which have been treated so far in the literature do not study multilinear relationships between variables. Such relationships are of a nonlinear nature. They take place whenever two or more than two univariate variables are the components of a multiple variable of order 2 or an order greater than 2. A multiple variable of order 2 is not a bivariate variable, and a multiple variable of an order greater than 2 is not a multivariate variable. A multiple variable allows for the construction of a tensor. The α -norm of this tensor gives rise to an aggregate measure of a multilinear nature. In particular, given a multiple variable of order 2, four regression lines can be estimated in the same subset of a two-dimensional linear space over R . How these four regression lines give rise to an aggregate measure of a multilinear nature is shown by this paper. In this research work, such a measure is an estimate concerning the expected return on a portfolio of financial assets. The metric notion of α -product is used to summarize the sampling units which are observed.

1. Introduction

Relationships between variables can be of a linear or nonlinear nature. In this paper, particular nonlinear relationships are made evident. Their intrinsic nature is multilinear. Multilinear relationships between variables take place whenever two or more than two univariate variables are the components of a multiple variable of order 2 or greater than 2. It is possible to write X 12 = {   1 X ,   2 X } , where   1 X and   2 X are the components of X 12 . If one writes X 123 = { 1 X ,   2 X ,   3 X } , then   1 X ,   2 X , and   3 X are the components of X 123 . Similarly, if one writes X 12 m = {   1 X ,   2 X ,   3 X , ,   m X } , then   1 X ,   2 X ,   3 X , …, and   m X are the components of X 12 m . Within this context, X 12 , X 123 , and X 12 m are multiple variables. Their components are univariate variables which influence each other. A multiple variable of order 2 is not a bivariate variable, and a multiple variable of order greater than 2 is not a multivariate variable. This paper will clarify why multilinear indices associated with a multiple variable of order 2 are based on 2 2 = 4 bivariate distributions of probability. Similarly, it will also be clarified why multilinear indices associated with a multiple variable of order m, where m > 2 is an integer, are based on m 2 bivariate distributions of probability. Each bivariate distribution of probability underlying a multilinear index is of a two-dimensional nature. Hence, it is possible to decompose it into two marginal probability distributions. This paper focuses on a multilinear approach of a Bayesian nature according to which more than two univariate variables are always studied in pairs. Since a multilinear and quadratic metric is used to study multilinear relationships between variables, various indices of a bivariate nature are firstly taken into account. This means that a disaggregation process first takes place. An aggregation process takes place second whenever a multilinear measure is shown. For these reasons, in this paper, given a multiple variable of order 2, four regression lines are estimated in the same subset of a two-dimensional linear space over R . The multilinear regression model introduced in this research work is different from those regression models which have been treated so far in the literature (see Aneiros et al. 2022; Ewald and Schneider 2020). The multilinear regression model handled in this paper allows us to make fair estimates based on the notion of prevision (see Cassese et al. 2020). Prevision consists of considering all the observed values drawn by a given population in order to distribute among them those nonnegative and finitely additive masses such that only absolutely inadmissible evaluations have to be excluded (see Angelini 2024b). Two dynamical phases are then involved with respect to the notion of prevision. They depend on a specific state of information and knowledge associated with a given individual. This state is intrinsically variable. The first phase gives rise to a closed convex set. This is because possible opinions about the evaluations of probability are infinite in number. Strictly speaking, if the observed values are put on a straight line, then a closed line segment appears in the first phase. Every point belonging to this closed line segment is a fair estimate of the parameter under consideration. The second phase gives rise to the choice of a point belonging to the closed line segment into account. This choice avoids that an extrapolation process takes place outside the closed line segment under consideration (see Berti et al. 2001). The notion of prevision is different from the one of prediction. Within this context, to make a prediction does not mean to handle barycenters of masses in the first phase and choosing one of them in the second one. To make a prediction means to make a prophecy. Hence, one out of the possible and observed values belonging to the set of possible alternatives can arbitrarily be chosen. It is then a prophecy. For this reason, one says that a prediction transforms the uncertainty into an artificial certainty (see Angelini 2024a). The distinction between prevision and prediction is not clear-cut in a non-specific language. Nevertheless, if one wants to identify an existing dichotomy, then this distinction is appropriate and necessary. There is therefore a fundamental difference: after knowing the specific outcome, a prediction can be said to be right or wrong in an objective way; conversely, whatever happens, nothing similar in any sense can be said about a prevision. The intrinsic nature of a prevision is then subjective.

The Objectives of This Paper

Given two variables   1 X and   2 X , if one writes   2 X = f (   1 X ) , then   2 X is the dependent or explained variable and   1 X is the independent or explanatory variable. Whenever a specific value of   1 X is predicted, the corresponding level associated with the other univariate variable   2 X can be one out of the observed values, or it can be found beyond them. If this specific value is found beyond observed values, then it can be found either to the left or right of them. The further one extrapolates beyond the observed values, the greater the risk one runs. Whenever the model under consideration is extrapolated far beyond the given sampling data, the result can become meaningless. In this paper, after studying four linear relationships between the two components,   1 X and   2 X , of a multiple variable of order 2 denoted by X 12 , one is always inclined to avoid the fact that aggregate estimates of a specific parameter take place by choosing, as their constituent elements, points outside the closed line segments belonging to the four regression lines, respectively. Such lines capture the given sampling units via the least squares criterion. For this reason, one refers oneself to the notion of prevision. Such a notion takes all given sampling units into account. It is possible to summarize them using the concept of α -product. Such a concept is of a metric nature. In Section 2, the two-variable multilinear regression model treated in this paper is conceptually and formally shown. In Section 3, a specific regression line is estimated. In Section 4, another regression line is estimated using the same observations in a different way. In Section 5, two regression lines, whose least squares residuals are intrinsically equal to zero, are handled. Section 6 shows that the notion of α -product is an essential tool of prevision. In Section 7, the bivariate correlation coefficient is obtained using an aggregate measure of a multilinear nature. This means that the bivariate correlation coefficient can intrinsically be referred to a multiple variable of order 2. Section 8 contains how the multilinear regression model treated in this paper can be used to make rational estimates related to the expected return on portfolios of financial assets. Section 9 shows that a tensor identifies a multiple variable. A fundamental difference between a multiple variable and a multivariate one is made evident. Moreover, the empirical validation of the proposed multilinear regression model is handled. Finally, Section 10 contains conclusions and some future perspectives.

2. The Two-Variable Multilinear Regression Model: Conceptual and Formal Properties

The two-variable multilinear regression model treated in this paper studies multilinear relationships between variables. X 12 is a multiple variable of order 2. Its components are denoted by   1 X and   2 X , respectively. One writes
X 12 = {   1 X ,   2 X } ,
where   1 X and   2 X are two univariate variables. The two-variable multilinear regression model handled in this research work is composed of four regression lines. They are depicted in the same subset of a two-dimensional linear space over R . They are studied in the same subset of the Cartesian plane. With respect to these four regression lines, two regression lines are estimated. The other two regression lines exactly coincide with the same 45-degree line. Such a model is formally given by
two-variable multilinear regression =   2 X = α + β ( 1 X ) ,   1 X = α * + β * ( 2 X ) ,   2 X =   1 X ,   1 X =   2 X .
Since   1 X and   2 X are the two components of X 12 , it is possible to observe that   1 X influences   2 X , and   2 X influences   1 X . There is no one-way causation. Hence,   1 X influences   2 X . Since there is feedback in the opposite direction,   2 X influences   1 X as well. Four different pairs of marginal variables are studied. They are (   1 X ,   1 X ) , (   1 X ,   2 X ) , (   2 X ,   1 X ) , and (   2 X ,   2 X ) . For this reason, a multilinear relationship between   1 X and   2 X is dealt with.

3. An Estimated Regression Line

Let
(   1 X i ,   2 X i ) ,
with i = 1 , , n , be a sample of n ordered pairs of observations associated with   1 X and   2 X , respectively. Given this sample, whose size is equal to n, the linear relationship between the two components of X 12 expressed by
  2 X = α + β (   1 X )
is first studied. It is possible to enlarge (4) by considering
  2 X = α + β (   1 X ) + u ,
where u is a stochastic variable which is normally distributed. One writes
  2 X i = α + β (   1 X i ) + u i ,
with i = 1 , , n , where it turns out to be
E ( u i ) = 0 , for all i , var ( u i ) = E ( u i 2 ) = σ 2 , for all i , cov ( u i u j ) = E ( u i u j ) = 0 , for i j .
The disturbances denoted by u i are independently and identically distributed, with zero mean and variance expressed by σ 2 .
The scatter of n observations is conventionally studied inside the first quadrant of a two-dimensional Cartesian coordinate system. An estimation of the linear relationship coinciding with (5) is given by the straight line expressed by
  2 X ^ = a + b (   1 X ) ,
where   2 X ^ indicates the height of the line at   1 X . If a and b are selected to minimize the residual sum of squares given by
RSS = i = 1 n (   2 e i )   2 = i = 1 n   2 X i a b (   1 X i ) 2 = f ( a , b ) ,
then the least squares principle takes place (see Watson 1967; Bartlett and Macdonald 1968). The necessary conditions for a stationary value of RSS are such that it is possible to write
a =   2 X ¯ b (   1 X ¯ ) ,
and
b = i = 1 n   1 x   i   2 x   i i = 1 n (   1 x   i )   2 ,
where one has
  1 x i =   1 X i   1 X ¯ ,
with i = 1 , , n , and
  2 x i =   2 X i   2 X ¯ ,
with i = 1 , , n . Once the sample means have been calculated, the data characterizing (11) are expressed using deviations from them.   1 X ¯ and   2 X ¯ denote the sample means of   1 X and   2 X , respectively.
The least squares line passes through the mean point given by (   1 X ¯ ,   2 X ¯ ) . The least squares residuals have zero correlation in the sample with the values of   1 X . They have also zero correlation with the values denoted by   2 X ^ . The decomposition of the sum of squares associated with the linear regression of   2 X on   1 X is written as
TSS = ESS + RSS ,
where one has
TSS = total sum of squared deviations in the   2 X variable , RSS = residual sum of squares from the regression of   2 X on   1 X , ESS = explained sum of squares from the regression of   2 X on   1 X .
Since it is possible to write
r 2 = ESS TSS ,
where the coefficient of determination given by r 2 is the square of the sample correlation coefficient expressed by r, (16) is the proportion of the   2 X variation attributable to the linear regression of   2 X on   1 X (see Fisher 1922). The limits of r are ± 1 . In the limiting case, the sample points all lie on a single straight line. The disturbance variance denoted by σ 2 cannot be estimated from a sample of u values denoted by u i , i = 1 , , n (see Liu et al. 2020). These values are unobservable. If one knows α and β , then they can be measured. Nevertheless, α and β are unknown. It follows that the disturbance variance can be estimated from the calculated residuals. It is accordingly possible to use the following estimation expressed by
  2 s 2 = i = 1 n (   2 e i )   2 n 2
to assess the disturbance variance.
Linear estimators of α and β are unbiased (see Puntanen and Styan 1989). One writes
E ( a ) = α
and
E ( b ) = β .
Their sampling variances are the smallest that can be achieved by any linear unbiased estimator. One writes
var ( a ) = σ 2 1 n + ( 1 X ¯ ) 2 i = 1 n (   1 x   i ) 2
and
var ( b ) = σ 2 i = 1 n (   1 x   i ) 2 .
The sampling distribution of the intercept term denoted by a is given by
a N α , σ 2 1 n + ( 1 X ¯ ) 2 i = 1 n (   1 x i ) 2 ,
whereas the sampling distribution of the least squares slope denoted by b is written as
b N β , σ 2 i = 1 n (   1 x i ) 2 .
Since σ 2 is unknown, it is necessary to use (17). It is an unbiased estimator because one observes
E i = 1 n (   2 e i )   2 n 2 = σ 2 .
One observes
i = 1 n (   2 e i )   2 σ 2 χ 2 ( n 2 ) .
Also, another fundamental result associated with inference procedures is that i = 1 n (   2 e i )   2 is distributed independently of f ( a , b ) . Tests on the intercept are based on the t distribution, so one writes
a α   2 s 1 n + ( 1 X ¯ ) 2 i = 1 n (   1 x   i ) 2 t ( n 2 ) .
Accordingly, a 100 ( 1 ϵ ) percent confidence interval for α is expressed by
a ± t ϵ / 2   2 s 1 n + ( 1 X ¯ ) 2 i = 1 n (   1 x   i ) 2 ,
whereas the null hypothesis given by H 0 :   α = α 0 is rejected at the 100 ϵ percent level of significance if one observes
a α 0   2 s 1 n + ( 1 X ¯ ) 2 i = 1 n (   1 x   i ) 2 > t ϵ / 2 .
Tests on b are based on the t distribution as well, so one writes
b β   2 s i = 1 n (   1 x   i ) 2 t ( n 2 ) .
A 100 ( 1 ϵ ) percent confidence interval for β is expressed by
b ± t ϵ / 2   2 s i = 1 n (   1 x   i ) 2 ,
whereas the null hypothesis given by H 0 :   β = β 0 is rejected at the 100 ϵ percent level of significance if one observes
b β 0   2 s i = 1 n (   1 x   i ) 2 > t ϵ / 2 .
Tests on σ 2 can be derived as well. It is possible to use a χ 2 variable to derive them.
The test for the significance of   1 X is given by H 0 :   β = 0 . It is known that there are three versions connected with it. First, this test can be set out in an analysis of variance framework using the F statistic expressed by
F = ESS / 1 RSS / ( n 2 ) .
On the other hand, it is also possible to base it on the sample correlation coefficient or on the regression slope using the t distribution in both cases.
Given an estimated regression line with intercept denoted by a and slope expressed by b, it is possible to focus own attention on a specific value given by   1 X 0 of the regressor variable. For example, if one is required to predict a 95 percent confidence interval for   2 X 0 , then one writes
a + b (   1 X 0 ) ± t 0.025   2 s 1 + 1 n + (   1 X 0   1 X ¯ ) 2 i = 1 n (   1 x   i ) 2 .
To test whether a new observation given by (   1 X 0 ,   2 X 0 ) comes from the structure generating the sample data, it is possible to compare this new observation with the confidence interval for   2 X 0 . If one is required to predict a 95 percent confidence interval for E (   2 X 0 ) , then it is possible to write
a + b (   1 X 0 ) ± t 0.025   2 s 1 n + (   1 X 0   1 X ¯ ) 2 i = 1 n (   1 x   i ) 2 .
In general, a sampling distribution shows the behavior of the estimators in repeated applications of the estimating formulae. A given sample provides a particular numerical estimation. Another sample from the same population will provide another numerical estimation. A sampling distribution shows the results that will be obtained for the estimators over the virtually infinite set of samples which could be drawn from the population into account. If one refers oneself to E (   2 X |   1 X ) = α + β (   1 X ) or E (   2 X |   1 X i ) = α + β (   1 X i ) , i = 1 , , n , then the parameters of interest are α , β , and σ 2 . They are parameters of interest of a conditional distribution. In this conditional distribution, the only source of variation from one hypothetical sample to another one is a variation in the stochastic disturbance denoted by u. This stochastic disturbance, together with the given values denoted by   1 X 1 , ,   1 X n , will determine the values denoted by   2 X 1 , ,   2 X n . It follows that the sample values of a, b, and   2 s 2 take place. Accordingly, analyzing   2 X conditional on   1 X deals with the   1 X 1 , ,   1 X n values viewed as fixed in repeated sampling. This is based on the implicit assumption that the marginal distribution for   1 X does not involve the parameters of interest. In other words, the marginal distribution for   1 X contains no information on α , β , and σ 2 .

4. Another Estimated Regression Line

This paper focuses on a regression model which is not only interested in viewing   1 X as the fixed regressor. This paper focuses on a model which is also interested in viewing   2 X as the fixed regressor. Accordingly, n ordered pairs of observations of the form given by
(   2 X i ,   1 X i ) ,
with i = 1 , , n , are now dealt with. One takes care to interchange   1 X i and   2 X i . One presently studies n ordered pairs of the same previous observations. Nevertheless, the first element of the previous n pairs of observations becomes the second one and vice versa. At present, one refers oneself to E (   1 X |   2 X ) = α * + β * (   2 X ) or E (   1 X |   2 X i ) = α * + β * (   2 X i ) , i = 1 , , n . If one refers oneself to E (   1 X |   2 X ) = α * + β * (   2 X ) or E (   1 X |   2 X i ) = α * + β * (   2 X i ) , i = 1 , , n , then the parameters of interest are α * , β * , and σ 2 . They are parameters of interest of another conditional distribution (see Gilio and Sanfilippo 2014). In this conditional distribution, the only source of variation from one hypothetical sample to another one is a variation in the stochastic disturbance. It is again denoted by u. This stochastic disturbance, together with the given values denoted by   2 X 1 , ,   2 X n , will determine the values denoted by   1 X 1 , ,   1 X n . It follows that the sample values of a * , b * , and   1 s 2 take place. Accordingly, analyzing   1 X conditional on   2 X deals with the   2 X 1 , ,   2 X n values viewed as fixed in repeated sampling. This is based on the implicit assumption that the marginal distribution for   2 X does not involve the parameters of interest. In other words, the marginal distribution for   2 X contains no information on α * , β * , and σ 2 . The linear relationship between the two components of X 12 expressed by
  1 X = α * + β * (   2 X )
is studied second. It is possible to enlarge (36) by considering
  1 X = α * + β * (   2 X ) + u ,
where u is a stochastic variable which is normally distributed. One writes
  1 X i = α * + β * (   2 X i ) + u i ,
with i = 1 , , n , where one observes
E ( u i ) = 0 , for all i , var ( u i ) = E ( u i 2 ) = σ 2 , for all i , cov ( u i u j ) = E ( u i u j ) = 0 , for i j .
The disturbances denoted by u i are independently and identically distributed, with zero mean and variance expressed by σ 2 .
An estimation of the linear relationship coinciding with (37) is given by the straight line expressed by
  1 X ^ = a * + b * (   2 X ) ,
where   1 X ^ indicates the value of the straight line which is observed on the horizontal axis at   2 X . If a * and b * are selected to minimize the residual sum of squares given by
RSS = i = 1 n (   1 e i )   2 = i = 1 n   1 X i a * b * (   2 X i ) 2 = f ( a * , b * ) ,
then the least squares principle takes place. The necessary conditions for a stationary value of RSS are such that it is possible to write
a * =   1 X ¯ b * (   2 X ¯ )
and
b * = i = 1 n   2 x   i   1 x   i i = 1 n (   2 x   i )   2 ,
where one has
  1 x i =   1 X i   1 X ¯ , i = 1 , , n
and
  2 x i =   2 X i   2 X ¯ , i = 1 , , n .
Once the sample means have been calculated, the data characterizing (42) are expressed using deviations from them.
The least squares line passes through the mean point given by (   1 X ¯ ,   2 X ¯ ) . The least squares residuals have zero correlation in the sample with the values of   2 X . They have also zero correlation with the values denoted by   1 X ^ . The decomposition of the sum of squares associated with the linear regression of   1 X on   2 X is written as
TSS = ESS + RSS ,
where one has
TSS = total sum of squared deviations in the   1 X variable , RSS = residual sum of squares from the regression of   1 X on   2 X , ESS = explained sum of squares from the regression of   1 X on   2 X .
Since it is possible to write
r 2 = ESS TSS ,
r 2 is the proportion of the   1 X variation attributable to the linear regression of   1 X on   2 X . The disturbance variance denoted by σ 2 cannot be estimated from a sample of u values denoted by u i , i = 1 , , n . These values are unobservable. If one knows α * and β * , then they can be measured. Unfortunately, α * and β * are unknown. It follows that the disturbance variance can be estimated from the calculated residuals. It is accordingly possible to use the following estimation expressed by
  1 s 2 = i = 1 n (   1 e i )   2 n 2
to assess the disturbance variance.
Linear estimators of α * and β * are unbiased. One writes
E ( a * ) = α *
and
E ( b * ) = β * .
Their sampling variances are the smallest that can be achieved by any linear unbiased estimator. One writes
var ( a * ) = σ 2 1 n + (   2 X ¯ ) 2 i = 1 n (   2 x   i ) 2
and
var ( b * ) = σ 2 i = 1 n (   2 x   i ) 2 .
The sampling distribution of the intercept term denoted by a * is given by
a * N α * , σ 2 1 n + (   2 X ¯ ) 2 i = 1 n (   2 x i ) 2 ,
whereas the sampling distribution of the least squares slope denoted by b * is written as
b * N β * , σ 2 i = 1 n (   2 x i ) 2 .
Since σ 2 is unknown, it is necessary to use (44). It is an unbiased estimator because one observes
E i = 1 n (   1 e i )   2 n 2 = σ 2 .
One observes
i = 1 n (   1 e i )   2 σ 2 χ 2 ( n 2 ) .
Also, another fundamental result associated with inference procedures is that i = 1 n (   1 e i )   2 is distributed independently of f ( a * , b * ) . Tests on the intercept are based on the t distribution, so one writes
a * α *   1 s 1 n + (   2 X ¯ ) 2 i = 1 n (   2 x   i ) 2 t ( n 2 ) .
Accordingly, a 100 ( 1 ϵ ) percent confidence interval for α * is expressed by
a * ± t ϵ / 2   1 s 1 n + (   2 X ¯ ) 2 i = 1 n (   2 x   i ) 2 ,
whereas the null hypothesis given by H 0 :   α * = α 0 * is rejected at the 100 ϵ percent level of significance if one observes
a * α 0 *   1 s 1 n + (   2 X ¯ ) 2 i = 1 n (   2 x   i ) 2 > t ϵ / 2 .
Tests on b * are based on the t distribution as well, so one writes
b * β *   1 s i = 1 n (   2 x   i ) 2 t ( n 2 ) .
A 100 ( 1 ϵ ) percent confidence interval for β * is expressed by
b * ± t ϵ / 2   1 s i = 1 n (   2 x   i ) 2 ,
whereas the null hypothesis given by H 0 :   β * = β 0 * is rejected at the 100 ϵ percent level of significance if one observes
b * β 0 *   1 s i = 1 n (   2 x   i ) 2 > t ϵ / 2 .
Tests on σ 2 can be derived as well. A χ 2 variable is used to derive them.
The test for the significance of   2 X is given by H 0 :   β * = 0 . There are three versions connected with it.
Given an estimated regression line with intercept denoted by a * and slope expressed by b * , it is possible to focus own attention on a specific value given by   2 X 0 of the regressor variable. For example, if one is required to predict a 95 percent confidence interval for   1 X 0 , then one writes
a * + b * (   2 X 0 ) ± t 0.025   1 s 1 + 1 n + (   2 X 0   2 X ¯ ) 2 i = 1 n (   2 x   i ) 2 .
To test whether a new observation given by (   1 X 0 ,   2 X 0 ) comes from the structure generating the sample data, it is possible to compare this new observation with the confidence interval for   1 X 0 . If one is required to predict a 95 percent confidence interval for E (   1 X 0 ) , then it is possible to write
a * + b * (   2 X 0 ) ± t 0.025   1 s 1 n + (   2 X 0   2 X ¯ ) 2 i = 1 n (   2 x   i ) 2 .

5. Two Regression Lines Whose Least Squares Residuals Are Equal to Zero

In addition to (3) and (35), this paper focuses on a regression model for which it is possible to consider n ordered pairs of observations of the form given by
(   1 X i ,   1 X i ) ,
with i = 1 , , n . This means that n ordered pairs of observations all lie on the 45-degree line denoted by
  2 X =   1 X .
Its intercept is exactly equal to 0, whereas its slope is exactly equal to 1. The least squares residuals are intrinsically equal to 0 (see del Pino 1989). Also, in this limiting case, the least squares estimation coincides with the maximum likelihood one (see McCullagh 1983; Nelder and Pregibon 1987).
In addition to (3), (35), and (61), this paper shows a regression model for which it is also possible to study n ordered pairs of observations of the form expressed by
(   2 X i ,   2 X i ) ,
with i = 1 , , n . This means that n ordered pairs of observations all lie on the 45-degree line denoted by
  1 X =   2 X .
Its intercept is exactly equal to 0, whereas its slope is exactly equal to 1. The least squares residuals are intrinsically equal to 0. In this limiting case, the least squares estimation coincides with the maximum likelihood one. From (62) and (64), it follows that it is true that the 45-degree line is the same graph depicted in the same subset of the Cartesian plane.

6. The Notion of α-Product as an Essential Tool of Prevision

With respect to the two estimated regression lines, one observes
  2 X =   2 X ^ +   2 e , with   2 e = 0 ,
and
  1 X =   1 X ^ +   1 e , with   1 e = 0 .
This means that two n-dimensional vectors denoted by
  2 X =   2 X 1   2 X 2   2 X n =   2 X 1   2 X 2   2 X n
and
  1 X =   1 X 1   1 X 2   1 X n =   1 X 1   1 X 2   1 X n
are considered together with two estimated regression lines. One also writes
  2 X ^ =   2 X 1 ^   2 X 2 ^   2 X n ^ =   2 X 1 ^   2 X 2 ^   2 X n ^
and
  1 X ^ =   1 X 1 ^   1 X 2 ^   1 X n ^ =   1 X 1 ^   1 X 2 ^   1 X n ^ .
These vectors belong to an n-dimensional linear space over R of a Euclidean nature. The last two vectors expressed by (69) and (70) are two n × 1 or 1 × n matrices involving regression residuals given by   2 e and   1 e , respectively. In fact, one observes
  2 X ^ =   2 X   2 e
and
  1 X ^ =   1 X   1 e .
In general, the scalar or inner product of two n-dimensional vectors is commutative, associative, and distributive. It may also be orthogonal, so the zero-product property does not hold. Since one can always consider the transpose of an n × 1 or 1 × n matrix, one writes
(   2 X ^ ) T   1 X =   2 X 1 ^   2 X 2 ^   2 X n ^   1 X 1   1 X 2   1 X n ,
where   2 X ^ and   1 X are two n × 1 matrices, or
  2 X ^ (   1 X ) T =   2 X 1 ^   2 X 2 ^   2 X n ^   1 X 1   1 X 2   1 X n ,
where   2 X ^ and   1 X are two 1 × n matrices. Anyway, one has
(   2 X ^ ) T   1 X =   2 X ^ (   1 X ) T .
The two-variable multilinear regression model treated in this research work uses the metric notion of α -product (see Angelini 2023). For example, the sampling quantities of the following system
  2 X = n a + b   1 X   1 X   2 X = a   1 X + b (   1 X ) 2
are five. They are n,   1 X ,   2 X ,   1 X   2 X , and (   1 X ) 2 . They characterize the normal equations for the linear regression of   2 X on   1 X . In this paper, such quantities are α -products. For example, the following Table 1 shows that the notion of α -product of two n-dimensional vectors is used to obtain   1 X . The following Table 2 shows that the notion of α -product of two n-dimensional vectors is used to obtain   1 X   2 X . Both two-way tables with n rows and n columns which have just been written use masses outside their main diagonal which are all equal to zero. In the first table, an n-dimensional vector, whose elements are all equal to 1, is used. With respect to the associative property of the notion of α -product, one can always use the constant given by 1 / n , where 1 / n R . This constant mathematically satisfies the conditions of coherence associated with the notion of prevision of a random entity (see von Rosen 1989; Coletti et al. 2016). From Table 1, it is possible to obtain
i = 1 n   1 X i n =   1 X ¯ ,
where the sum of positive and equiprobable masses associated with   1 X 1 , …,   1 X n is coherently equal to 1. On the other hand, if all deviations from   1 X ¯ with respect to the first vector of Table 2 are considered together with all deviations from   2 X ¯ with respect to the second one of the same table, then it is possible to obtain the expression given by
cov (   1 X ,   2 X ) = i = 1 n   1 x   i   2 x   i n .
Let (   1 X ^ ,   1 X ^ ) , (   1 X ,   2 X ^ ) , (   2 X ,   1 X ^ ) , and (   2 X ^ ,   2 X ^ ) be four ordered pairs of variables. They identify four two-way tables, where each of them has n rows and n columns. The n nonnegative masses appearing on the main diagonal of each two-way table are such that their sum is always equal to 1. All observed values are possible. Since there are no degrees of possibility, one can attribute to the various observed values a greater or lesser degree of a specific factor of a psychological nature expressing a more or less strong propensity to expect that a given value rather than others will turn out to be true at the right time (see Chen and Escobar-Anel 2021; Flores et al. 2021). The two-variable multilinear regression model treated in this paper can make coherent previsions. After establishing how all the n nonnegative masses appearing on the main diagonal of each two-way table are, the α -norm of a tensor of order 2 is handled (see Maturo and Angelini 2023). Such a norm coincides with the determinant of a square matrix of order 2, whose elements are all α -products. One writes
E (   1 X ^   1 X ^ ) E (   1 X   2 X ^ ) E (   2 X   1 X ^ ) E (   2 X ^   2 X ^ ) ,
so it is possible to obtain
E ( X ^ 12 ) = P ( X ^ 12 ) = E (   1 X ^   1 X ^ ) E (   1 X   2 X ^ ) E (   2 X   1 X ^ ) E (   2 X ^   2 X ^ ) = P (   1 X ^   1 X ^ ) P (   1 X   2 X ^ ) P (   2 X   1 X ^ ) P (   2 X ^   2 X ^ ) ,
where E ( X ^ 12 ) = P ( X ^ 12 ) is a real number. The same symbol P signifies the unique notion which, in general, is called prevision or mathematical expectation. In the case of single events, it is also called probability. The α -products characterizing the two elements of the main diagonal of (79) are associated with the 45-degree line, whereas the α -products characterizing the two elements of the antidiagonal of the same square matrix are connected with the two estimated regression lines. E (   1 X ^   1 X ^ ) , E (   1 X   2 X ^ ) , E (   2 X   1 X ^ ) , and E (   2 X ^   2 X ^ ) are all bilinear previsions obtained by considering n nonnegative and finitely additive masses associated with n points. Each bilinear prevision is always decomposed into two linear previsions. This means that points on a specific regression line are always studied in the Cartesian plane. In the first phase, each mass of n masses can take infinite values between 0 and 1, endpoints included, into account. Hence, a closed convex set is handled with respect to each regression line. It is a closed line segment that is bounded by the two extreme points which are studied on each regression line. Given a two-way table, all the masses outside its main diagonal are always equal to zero. Straightforward cases are the following:
Example 1. 
If y = x is a regression line, then the following Table 3 shows how a bivariate (two-dimensional) probability distribution can be decomposed. A point belonging to y = x is given by ( 3.3 , 3.3 ) . A closed line segment appears on y = x in the first phase. Its endpoints are ( 0 , 0 ) and ( 4 , 4 ) . The second phase depends on the state of information and knowledge associated with a given individual. On the other hand, if y = 2 x is an estimated regression line, then the following Table 4 shows that a point belonging to y = 2 x is expressed by ( 2.6 , 5.2 ) . A closed line segment appears on y = 2 x in the first phase. Its endpoints are ( 0 , 0 ) and ( 3 , 6 ) . The choice of a point among infinite points on the estimated regression line depends on the state of information and knowledge associated with a given individual. This choice is absolutely rational.

7. The Bivariate Correlation Coefficient Obtained Using an Aggregate Measure of a Multilinear Nature

The two-variable multilinear regression model treated in this paper focuses on a multiple variable of order 2. This section shows that the sample correlation coefficient is intrinsically associated with a multiple variable of order 2. Such a coefficient is written in the following form given by
r = r 12 =   1 x   2 x (   1 x ) 2 (   2 x ) 2 ,
where deviations from two mean values are considered (see Yule 1897). The population correlation coefficient is always associated with a multiple variable of order 2 in the same way as the sample correlation coefficient. If the notion of α -product of two n-dimensional vectors expressed using deviations is taken into account, then it is possible to write the sample correlation coefficient in the following form given by
r = r 12 =   ( 1 )   d       ( 2 )   d       ( 1 )   d     α   ( 2 )   d     α ,
where it turns out to be
  ( 1 )   d       ( 1 )   d     =   ( 1 )   d     α 2 =   ( 1 )   d     α
and
  ( 2 )   d       ( 2 )   d     =   ( 2 )   d     α 2 =   ( 2 )   d     α .
The symbol given by ⊙ identifies the notion of α -product of two vectors. The sample correlation coefficient geometrically depends on the size of the angle between   ( 1 )   d     and   ( 2 )   d     , where   ( 1 )   d     and   ( 2 )   d     are two n-dimensional vectors located at the origin of an n-dimensional linear space over R of a Euclidean nature. Since it is possible to write
1 1   12   d     α 2   12   d ^     α 2 1 / 2 1 ,
where one has
  12   d α 2 =   ( 1 )   d     α 2   ( 1 )   d       ( 2 )   d       ( 2 )   d       ( 1 )   d       ( 2 )   d     α 2
and
  12   d ^ α 2 =   ( 1 )   d     α 2 0 0   ( 2 )   d     α 2 ,
the expression within the parentheses of (85) coincides with the sample correlation coefficient denoted by r 12 . Such a coefficient is then proved to be intrinsically referred to X 12 = {   1 X ,   2 X } . The expression within the parentheses of (85) has been obtained using tensors of order 2 referred to a multiple variable of order 2. The sample correlation coefficient denoted by r 12 = r (   1 X ,   2 X ) is invariant with respect to rotations and translations. One writes
r a (   1 X ) + b , c (   2 X ) + d = r (   1 X ,   2 X ) ,
with a , c 0 . It is possible to verify (88) using a multilinear and quadratic metric. Moreover, one observes
r (   1 X ,   2 X ) = r (   2 X ,   1 X )
and
r (   1 X ,   1 X ) = r (   2 X ,   2 X ) = 1 .
Whenever a correlation between   1 X and   2 X is studied, it is possible to express the underlying multilinear relationship between   1 X and   2 X through the following determinant
r (   1 X ,   1 X ) r (   1 X ,   2 X ) r (   2 X ,   1 X ) r (   2 X ,   2 X ) .
However, in this case, it does not make sense that the calculation of (91) takes place. On the other hand, it is known that one of the most meaningful relationships constructed from a finite set of elements is a linear combination of basis vectors of a given linear space. Nevertheless, given some vectors belonging to the same linear space, an actual linear combination of them does not always make sense in practice: this observation is therefore conceptually the same as the one above related to (91).

8. The Two-Variable Multilinear Regression Model Used to Study Portfolios of Financial Assets: Fair Estimations

An n-dimensional linear space over R of a Euclidean nature is denoted by R n .   1 X and   2 X are elements of R n , where R n is isomorphic to R n . An n-dimensional affine space is expressed by R n . The two elements denoted by   1 X and   2 X are supposed to be linearly independent. They identify a linear subspace of R n . Its dimension is equal to 2. One observes a reduction of dimension in this way (see Yuan et al. 2007). With respect to   1 X , all collinear vectors referred to it are considered. Their number is infinite, so a straight line of R n is established. This line is a linear subspace of R n . Its dimension is equal to 1. It is also a linear subspace of R n . It is possible to prove in the form of a theorem that there is a one-to-one correspondence between the elements of a sheaf of parallel hyperplanes embedded in R n and the points of intersection of them with a straight line of R n . This line is orthogonal to all parallel hyperplanes taken into account. The points of intersection of all these parallel hyperplanes embedded in R n with a straight line of R n are real numbers transferred on a one-dimensional straight line, on which an origin, a unit of length, and an orientation are established. They coincide with all the elements of   1 X . The same is true with respect to all the elements of   2 X . Two mutually orthogonal one-dimensional straight lines, on which an origin, a same unit of length, and an orientation are established, identify a two-dimensional Cartesian coordinate system. Ordered pairs of sampling units referring to two univariate variables denoted by   1 X and   2 X can be studied inside the first quadrant of a two-dimensional Cartesian coordinate system. This is because it is always possible to write
  1 X =   1 X   1   1 X   2 ,
with
  1 X   1 = 0   1 X ,
and
  1 X   2 = | 0   1 X | .
Moreover, it is always possible to write
  2 X =   2 X   1   2 X   2 ,
with
  2 X   1 = 0   2 X ,
and
  2 X   2 = | 0   2 X | .
In general, one has
x y = max ( x , y ) ,
x y = min ( x , y ) ,
and
x ˜ = 1 x ,
for any x and y that are real numbers.
The two-variable multilinear regression model treated in this paper shows that the real world cannot be analyzed in terms of a collection of bivariate or multivariate relationships only. The real world can also be analyzed in terms of multilinear relationships between variables (see Angelini and Maturo 2022). Surveys and technical discussions of the relationships between variables, which are studied in pairs, steadily take place. Various relationships between variables studied in pairs can be treated. Different pairs of variables are analyzed whenever a multilinear approach is used. This approach uses a multilinear and quadratic metric. For this reason, different pairs of variables have to be considered. Even though multilinear relationships between variables can be present whenever real problems are studied, one does not leave the two-dimensional diagrams behind. Bivariate relationships between variables are significant in themselves. The mathematical and statistical tools developed for them are fundamental building blocks for the analysis of multilinear situations characterizing real problems (see Bonat and Jørgensen 2016). X 12 = { 1 X ,   2 X } can identify a two-financial asset portfolio. Its components can be two normally distributed variables. In other words, a bivariate normal population is considered. This means, among other things, that the   1 X values identifying observed returns on   1 X are normally distributed and so are the   2 X values identifying observed returns on   2 X . A multiple variable of order 2 is associated with X ^ 12 , where X ^ 12 identifies 2 2 = 4 regression lines which are studied inside the same subset of a two-dimensional linear space over R . In this paper, a two-financial asset portfolio is studied as a multiple random good of order 2. This matter can be extended, so
X 123 = { 1 X ,   2 X ,   3 X }
can identify a three-financial asset portfolio. Its components are three normally distributed variables, so a multivariate normal population is considered. This means that the   1 X values identifying observed returns on   1 X are normally distributed. The   2 X values identifying observed returns on   2 X are normally distributed and so are the   3 X values identifying observed returns on   3 X . A multiple variable of order 3 is associated with X ^ 123 , where X ^ 123 identifies 3 2 = 9 regression lines. They are studied inside the same subset of a two-dimensional linear space over R . A quadratic metric is always used (see Markowitz 1952, 1956; Cont 2001). If X 12 = { 1 X = X ,   2 X = Y } is a continuous and multiple variable of order 2, whose components are two continuous and univariate variables, then its expected return is given by the following determinant
E ( X 2 ) = + x 2 f ( x ) d x E ( X Y ) = + + x y f ( x , y ) d x d y E ( Y X ) = + + y x f ( y , x ) d y d x E ( Y 2 ) = + y 2 f ( y ) d y ,
where f ( x ) and f ( y ) are density functions of X and Y, whereas f ( x , y ) is a bivariate probability density function like f ( y , x ) . This determinant has the same structure as the one characterizing the α -norm of a tensor of order 2 connected with sampling quantities and their regression residuals. All elements of this determinant are integrals, so it is necessary to verify they exist. Similarly, if X 123 = { 1 X = X ,   2 X = Y ,   3 X = Z } is a continuous and multiple variable of order 3, whose components are three continuous and univariate variables, then its expected return is given by the following determinant
E ( X 2 ) = + x 2 f ( x ) d x E ( X Y ) = + + x y f ( x , y ) d x d y E ( X Z ) = + + x z f ( x , z ) d x d z E ( Y X ) = + + y x f ( y , x ) d y d x E ( Y 2 ) = + y 2 f ( y ) d y E ( Y Z ) = + + y z f ( y , z ) d y d z E ( Z X ) = + + z x f ( z , x ) d z d x E ( Z Y ) = + + z y f ( z , y ) d z d y E ( Z 2 ) = + z 2 f ( z ) d z ,
where f ( z ) is a density function of Z, whereas f ( x , z ) and f ( y , z ) are bivariate probability density functions like f ( z , x ) and f ( z , y ) . It is also possible to study an m-financial asset portfolio denoted by
X 12 m = {   1 X ,   2 X ,   3 X , ,   m X } ,
where m > 3 is an integer. In particular, it is possible to treat its riskiness instead of its expected return (see Cepeda and Gamerman 2000; Angelini and Maturo 2023). Conceptually, if one thinks of the case of the barycenter of a solid body, then it is possible to understand why to set up a determinant to obtain a multilinear index related to a multiple variable is meaningful. The center of mass of a distribution of mass can always be established. Conversely, it is never possible in practice to know the exact distribution of mass related to a specific solid body. Similarly, to know the exact probability distribution related to a multiple variable is of little interest in many problems, but to know its mathematical expectation is essential.
The two-variable multilinear regression model handled in this research work extends other models widely studied in the literature (see Nelder and Wedderburn 1972). It is not in conflict with them (see Yuan and Lin 2006). Since the two-variable multilinear regression model is conceptually and mathematically associated with the bivariate correlation coefficient, the null hypothesis given by
H 0 :   ρ = 0 ,
where ρ is the population correlation coefficient, is rejected if the null hypothesis given by
H 0 :   β = 0 ,
or
H 0 :   β * = 0
is rejected. To reject β = 0 or β * = 0 implies that it is absolutely correct to reject ρ = 0 . It follows that there is a correlation between the two components of the two-financial asset portfolio under consideration.

9. Aggregate Measures: Multiple Random Variables and Their Identification through Tensors

If one studies a multiple random variable, then one obtains an aggregate measure. This is because one has the possibility of constructing a tensor, after which one has the possibility of calculating the α -norm of this tensor by applying the techniques of multilinearity. In this way, the logical aspects of probability calculus are studied in deep. Also, they give a heuristic contribution. These aspects must be kept unconnected from the empirical ones. This approach is of a Bayesian nature. The probability distributions which are involved can first be of a nonparametric kind. In a multiple variable, the role played by the marginal variables characterized by their marginal probability distributions is essential. Within this context, the support of each marginal variable is a partition of events containing a finite number of them. An event is not a measurable set, but it is a possible alternative expressed by a real number to which a specific probability is associated. The notion of probability is not undefined, much the same as point and line are undefined in geometry. The notion of event is infinitely subdivisible. However, a subdivision stops when one studies the Cartesian product of two sets, where each set contains a finite number of possible alternatives. The Cartesian product of two sets is studied because otherwise things become unnecessarily complicated. The Cartesian product of two sets of possible alternatives identifies a bivariate variable. Nevertheless, the choice of the ordered pair of marginal variables identifying a bivariate variable is arbitrary. For this reason, it is appropriate to free oneself from the order in which the marginal variables are taken into account. Given a multiple random variable of order n containing n marginal variables written as
X 12 n = {   1 X ,   2 X ,   3 X , ,   n X } ,
it is possible to construct the following tensor expressed by
P (   1 X   1 X ) P (   1 X   2 X ) P (   1 X   n X ) P (   2 X   1 X ) P (   2 X   2 X ) P (   2 X   n X ) P (   n X   1 X ) P (   n X   2 X ) P (   n X   n X ) .
Each element of the square matrix of order n given by (109) is a component of a tensor of order n. The components of this tensor are symmetric. Each element of (109) is an α -product obtained using an affine tensor of order 2. This tensor identifies the joint probabilities characterizing each bivariate probability distribution associated with a bivariate variable. Each bivariate variable is intrinsically composed of two marginal variables. The possible alternatives for each marginal variable are the components of a vector, so each bivariate probability distribution is summarized through an α -product structurally referred to the tensor product of two vectors. The tensor product of two vectors is not a scalar or inner product. The tensor product of two vectors is not commutative. Hence, n 2 ordered pairs of marginal variables are handled. The α -norm of the above-written tensor is given by
P ( X 12 n ) = P (   1 X   1 X ) P (   1 X   2 X ) P (   1 X   n X ) P (   2 X   1 X ) P (   2 X   2 X ) P (   2 X   n X ) P (   n X   1 X ) P (   n X   2 X ) P (   n X   n X ) .
It is the determinant of the square matrix of order n given by (109). There is no loss of information. On the contrary, there is a gain of information. This is because an aggregate measure compatible with objects of a homogeneous nature is obtained. On the other hand, an aggregate measure expressing the riskiness of a finite set of financial assets is given by
var ( X 12 n ) = 1 2     2 Δ 2 ( 1 X ) cov ( 1 X ,   2 X ) cov ( 1 X ,   n X ) cov ( 2 X ,   1 X ) 1 2     2 Δ 2 ( 2 X ) cov ( 2 X ,   n X ) cov ( n X ,   1 X ) cov ( n X ,   2 X ) 1 2     2 Δ 2 ( n X ) ,
where the mean quadratic difference of   i X is denoted by     2 Δ ( i X ) , i = 1 , , n , the square of the mean quadratic difference of   i X is denoted by     2 Δ 2 ( i X ) , i = 1 , , n , and one writes     2 Δ 2 ( i X ) = 2 σ   i X 2 , i = 1 , , n . It follows that it is possible to observe
var ( X 12 n ) = var ( 1 X ) cov ( 1 X ,   2 X ) cov ( 1 X ,   n X ) cov ( 2 X ,   1 X ) var ( 2 X ) cov ( 2 X ,   n X ) cov ( n X ,   1 X ) cov ( n X ,   2 X ) var ( n X ) .
Aggregate measures are essential in many applications of an economic nature. For instance, the Sharpe ratio can be obtained using a multilinear approach. This approach gives rise to multilinear measures. If r f is the risk-free asset paying a fixed rate of return, then the Sharpe ratio is given by
SR = P ( X 12 n ) r f var ( X 12 n ) .
The Sharpe ratio measures how risk and return can be traded off in making portfolio choices. One assumes P ( X 12 n ) > r f . Furthermore, the beta of a given stock i is expressed by
β i = cov ( r i , r m ) var ( r m ) ,
so aggregate measures obtained according to a multilinear approach can operationally be associated with the capital asset pricing model (CAPM). This model has many applications in the study of financial markets. The market expected return r m is expressed by (110). The mean-variance model based on mean return and standard deviation of return can use aggregate measures which are perfectly compatible with the satisfaction of the optimization principle by a given individual. The price of risk is then given by
p = r m r f σ m ,
where r m and σ m are always determinable within this context, and it represents the positive slope of the budget line. This is because one assumes that r m > r f . It is possible to identify a part of this budget line that is bounded by two distinct endpoints. This part is a closed line segment containing every point on the budget line that is between its endpoints. This paper focuses on a specific aspect: there are always infinitely many points for which the optimization principle can be satisfied by a given individual. They are points belonging to a closed convex set. It is known that the mean-variance model is a reasonable approximation to the expected utility model. Even the expected utility model can use aggregate measures obtained according to a multilinear approach. This means that Jensen’s inequality can be extended. Studying n marginal variables, which are the components of X 12 n , the certain gain, which is considered equivalent to X 12 n by a given individual who is risk neutral, is expressed by x 12 n = P ( X 12 n ) . His utility function is the 45-degree line. For a risk-loving individual, the random gain denoted by X 12 n is preferred to a certain gain denoted by x 12 n , so x 12 n is greater than P ( X 12 n ) on the x-axis. His utility function is therefore convex. For a risk-averse individual, the random gain X 12 n is not preferred to a certain gain x 12 n , so x 12 n is less than P ( X 12 n ) on the x-axis. His utility function is then concave.

9.1. A Multiple Random Variable: Emphasis on Marginal Variables

Check the following:
Example 2. 
Given X 123 = {   1 X ,   2 X ,   3 X } , the corresponding tensor is a square matrix of order 3. It is then possible to consider 3 2 = 9 ordered pairs of marginal variables. Whenever 3 2 = 9 ordered pairs of marginal variables are studied, each marginal probability distribution is summarized in such a way that one obtains a marginal measure. After summarizing, each marginal measure never changes. The joint probabilities can freely be chosen by a given individual. The only constraint for them is that the mean value associated with each marginal probability distribution does not change. The following Table 5 shows how the mean value of   1 X and   2 X can be. The following Table 6 shows how the mean value of   1 X and   3 X can be. The following Table 7 shows how the mean value of   2 X and   3 X can be. Given the mean values of   1 X ,   2 X , and   3 X , it is theoretically possible to consider all deviations from them, respectively. Thus, if the number of possible alternatives for   1 X ,   2 X , and   3 X is not small, then the joint probabilities can be chosen in such a way that the correlation coefficient is equal to + 1 , or 0, or 1 . A value of + 1 implies that a marginal variable increases as another one increases, whereas a value of 1 implies that a marginal variable increases while another one decreases. A value of 0 implies that there is no linear dependency between the corresponding marginal variables. If a marginal variable is studied together with itself, then only the marginal probabilities of its distribution appear as joint ones. Whenever the joint probabilities are freely chosen by a given individual, they do not change. The joint probabilities can be chosen based on the state of information and knowledge associated with a given individual.
It is necessary to handle errors associated with the notion of prevision. Natural or intrinsic errors connected with this notion appear whenever deviations from mean values take place. These errors are not unwelcome. They identify a parametric probability distribution. They identify a normal distribution characterized by specific parameters. Conversely, logical errors appear whenever objective conditions of coherence are not satisfied. Such conditions are not satisfied when and only when one chooses a point that does not belong to the corresponding closed convex set. These errors are absolutely unwelcome. It is possible to note the following:
Remark 1. 
The empirical validation of the multilinear regression model treated in this paper can easily be shown. This is because the α-norm of a tensor extends the properties of the barycenter of nonnegative and finitely additive masses. It is known that such properties are given by two distinct indices: the first index is associated with an equilibrium that is stable, the second one appears when the moment of inertia is minimum. Such properties are satisfied when and only when each component of the tensor into account is coherently expressed in a quantitative form obtained from a specific α-product. The α-norm of the tensor of order 3 related to Example 2 is given by
P ( X 12 3 ) = P (   1 X   1 X ) P (   1 X   2 X ) P (   1 X   3 X ) P (   2 X   1 X ) P (   2 X   2 X ) P (   2 X   3 X ) P (   3 X   1 X ) P (   3 X   2 X ) P (   3 X   3 X ) .
If each element of the following square matrix of order 3 given by
P (   1 X   1 X ) P (   1 X   2 X ) P (   1 X   3 X ) P (   2 X   1 X ) P (   2 X   2 X ) P (   2 X   3 X ) P (   3 X   1 X ) P (   3 X   2 X ) P (   3 X   3 X )
is coherent, then P ( X 12 3 ) is a parameter of a coherent nature. This means that P ( X 12 3 ) is a point belonging to the union of different closed convex sets. Each closed convex set is a part of a one-dimensional straight line that is bounded by two distinct endpoints. The product of a possible value for a marginal variable and a possible value for another one is a point of a one-dimensional straight line. This validation holds regardless of the particular masses that may be chosen by a given individual. The reader can verify this matter. As a consequence, the validation of the multilinear regression model treated in this paper is of a logical nature as well.
Remark 2. 
If the number of the possible values for two marginal variables is not the same, then this number can become the same. For instance, let
2 3 4 5 , 0.1 0.4 0.2 0.3
and
1 7 8 , 0.2 0.6 0.2
be two ordered pairs of vectors. The second vector of each ordered pair identifies the corresponding probabilities. If one writes
2 3 4 5 , 0.1 0.4 0.2 0.3
and
0 1 7 8 , 0 0.2 0.6 0.2 ,
then the number of the possible values for the two marginal variables under consideration becomes the same. The possible values for the two marginal variables into account identify the first vector of each ordered pair.

9.2. Difference between a Multiple Random Variable and a Multivariate One

There is a fundamental logical difference between a multiple random variable and a multivariate one. For instance, if a multivariate random variable is composed of three marginal probability distributions, then the probability distribution of this variable is of a three-dimensional nature. It is then possible to consider three axes belonging to R 3 . They are pairwise orthogonal. Each axis of them is the support of a marginal variable. Each marginal probability distribution which is summarized gives rise to a real number. Each real number associated with a marginal variable is represented on an axis belonging to R 3 . Each axis of R 3 is a linear subspace of it. Its dimension is equal to 1. If the probability distribution of a multivariate random variable composed of three marginal probability distributions is summarized, then an ordered triplet of real numbers belonging to R 3 has to be studied. This is because the probability distribution of a multivariate random variable composed of three marginal probability distributions is of a three-dimensional nature. Whenever the probability distribution of a multivariate random variable composed of three marginal probability distributions is summarized and a real number is handled, only empirical aspects are taken into account. Nevertheless, this paper says that it is necessary to consider logical aspects as well. For this reason, the previous real number is never an aggregate measure. A multiple random variable of order 3 logically gives rise to a tensor of order 3 which is naturally an element of a linear space over R . Thus, 3 2 = 9 ordered pairs of marginal variables are dealt with. Some ordered pairs out of 3 2 = 9 give rise to a bivariate variable such that the corresponding bivariate probability distribution is summarized inside a linear subspace of R 3 of a two-dimensional nature. This linear subspace is generated by two linear subspaces of R 3 of a one-dimensional nature. Conversely, three ordered pairs of marginal variables out of 3 2 = 9 give rise to a bivariate variable such that the corresponding bivariate probability distribution is summarized inside a linear subspace of R 3 of a one-dimensional nature. It is one of the three axes of R 3 : only the squares of the possible values for each marginal variable are taken into account, so there is no loss of information. On the other hand, the same is true if a multiple random variable is of order n, where n is an integer and one has n = 2 or n > 3 , and the probability distribution of a multivariate random variable is of an n-dimensional nature. Within this context, the theorem mentioned at the beginning of Section 8 does not come into play.

10. Conclusions and Some Future Perspectives

Two dynamical phases depending on the state of information and knowledge associated with a given individual always identify the notion of prevision. This notion expresses the judgment of a given individual at a given time. A given individual can criticize his previous judgment. He can therefore verify whether he forgot or underestimated or overestimated some circumstances. Such circumstances could have led him to a more accurate prevision if he had exercised better thought. Nevertheless, if his previous judgment translated into quantitative terms is an element of a closed convex set, then his past prevision is never wrong. There are only natural or intrinsic errors. New pieces of information can be used to make different previsions for more or less similar future cases. It is absolutely wrong to criticize a prevision based on a specific state of information and knowledge by referring to a different state of information and knowledge. In this paper, closed convex sets play an essential role. Closed convex sets are closed line segments. Each line segment is a part of a line regression. Fair estimations are fair previsions. They are studied inside linear spaces over R provided with a quadratic metric. Only their dimension can be different. Since it is possible to show that an invariant or intrinsic metric can be used, only the affine properties make sense. All indices which are obtained inside linear spaces over R provided with a quadratic metric do not depend on the arbitrary choice of the coordinate system. They are therefore of an intrinsic nature. This paper shows that obtaining aggregate measures is meaningful. For this reason, different applications of an economic nature are mentioned. Thus, if a given individual is indifferent to the exchange of   1 X for P (   1 X ) and of   2 X for P (   2 X ) , then the same individual has to be indifferent to the exchange of X 12 for P ( X 12 ) , where P ( X 12 ) is an aggregate measure. P ( X 12 ) is a real number obtained from four bivariate probability distributions. They are all summarized and decomposed. In this paper, a bivariate probability distribution is of a two-dimensional nature. A bivariate probability distribution that is summarized is not properly associated with a real number. A bivariate probability distribution that is summarized is intrinsically composed of two marginal probability distributions that are summarized. A bivariate probability distribution that is summarized can then be associated with an ordered pair of real numbers. Logical aspects of probability calculus are studied in depth. Also, they give a heuristic contribution. To determine an aggregate measure is crucial in connection with the optimization of decisions of an economic nature. Such decisions usually depend on the subjective preferences associated with a given individual, who can be interested in two or more than two components of a multiple variable. These components are studied using the techniques of multilinearity. Among other things that can represent future perspectives, it is possible to prove that closed convex sets are of an essential importance in connection with stochastic models related to fair estimations. Such models are intrinsically based on a Bayesian reinterpretation of the central limit theorem. Moreover, it is possible to make evident important aspects of time series through a reinterpretation of principal component analysis based on the techniques of multilinearity.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The author can confirm that all relevant data are included in the article.

Conflicts of Interest

The author declares that they have no conflicts of interest.

References

  1. Aneiros, Germán, Silvia Novo, and Philippe Vieu. 2022. Variable selection in functional regression models: A review. Journal of Multivariate Analysis 188: 104871. [Google Scholar] [CrossRef]
  2. Angelini, Pierpaolo. 2023. Probability spaces identifying ordinal and cardinal utilities in problems of an economic nature: New issues and perspectives. Mathematics 11: 4280. [Google Scholar] [CrossRef]
  3. Angelini, Pierpaolo. 2024a. Financial decisions based on zero-sum games: New conceptual and mathematical outcomes. International Journal of Financial Studies 12: 56. [Google Scholar] [CrossRef]
  4. Angelini, Pierpaolo. 2024b. Invariance of the mathematical expectation of a random quantity and its consequences. Risks 12: 14. [Google Scholar] [CrossRef]
  5. Angelini, Pierpaolo, and Fabrizio Maturo. 2022. The price of risk based on multilinear measures. International Review of Economics and Finance 81: 39–57. [Google Scholar] [CrossRef]
  6. Angelini, Pierpaolo, and Fabrizio Maturo. 2023. Tensors associated with mean quadratic differences explaining the riskiness of portfolios of financial assets. Journal of Risk and Financial Management 16: 369. [Google Scholar] [CrossRef]
  7. Bartlett, Maurice S., and Peter D. M. Macdonald. 1968. “Least-squares” estimation of distribution mixtures. Nature 217: 195–96. [Google Scholar] [CrossRef]
  8. Berti, Patrizia, Eugenio Regazzini, and Pietro Rigo. 2001. Strong previsions of random elements. Statistical Methods and Applications (Journal of the Italian Statistical Society) 10: 11–28. [Google Scholar] [CrossRef]
  9. Bonat, Wagner Hugo, and Bent Jørgensen. 2016. Multivariate covariance generalized linear models. Journal of the Royal Statistical Society, Series C (Applied Statistics) 65: 649–75. [Google Scholar] [CrossRef]
  10. Cassese, Gianluca, Pietro Rigo, and Barbara Vantaggi. 2020. A special issue on the mathematics of subjective probability. Decisions in Economics and Finance 43: 1–2. [Google Scholar] [CrossRef]
  11. Cepeda, Edilberto, and Dani Gamerman. 2000. Bayesian modeling of variance heterogeneity in normal regression models. Brazilian Journal of Probability and Statistics 14: 207–21. [Google Scholar]
  12. Chen, Junhe, and Marcos Escobar-Anel. 2021. Model uncertainty on commodity portfolios, the role of convenience yield. Annals of Finance 17: 501–28. [Google Scholar] [CrossRef]
  13. Coletti, Giulianella, Davide Petturiti, and Barbara Vantaggi. 2016. When upper conditional probabilities are conditional possibility measures. Fuzzy Sets and Systems 304: 45–64. [Google Scholar] [CrossRef]
  14. Cont, Rama. 2001. Empirical properties of asset returns: Stylized facts and statistical issues. Quantitative Finance 1: 223–36. [Google Scholar] [CrossRef]
  15. del Pino, Guido. 1989. The unifying role of iterative generalized least squares in statistical algorithms. Statistical Science 4: 394–403. [Google Scholar]
  16. Ewald, Karl, and Ulrike Schneider. 2020. On the distribution, model selection properties and uniqueness of the Lasso estimator in low and high dimensions. Electronic Journal of Statistics 14: 944–69. [Google Scholar] [CrossRef]
  17. Fisher, Ronald A. 1922. The goodness of fit of regression formulae, and the distribution of regression coefficients. Journal of the Royal Statistical Society 85: 597–612. [Google Scholar] [CrossRef]
  18. Flores, Brandon, Blessing Ofori-Atta, and Andrey Sarantsev. 2021. A stock market model based on CAPM and market size. Annals of Finance 17: 405–24. [Google Scholar] [CrossRef]
  19. Gilio, Angelo, and Giuseppe Sanfilippo. 2014. Conditional random quantities and compounds of conditionals. Studia logica 102: 709–29. [Google Scholar] [CrossRef]
  20. Liu, Xu, Shurong Zheng, and Xingdong Feng. 2020. Estimation of error variance via ridge regression. Biometrika 107: 481–88. [Google Scholar] [CrossRef]
  21. Markowitz, Harry M. 1952. The utility of wealth. Journal of Political Economy 60: 151–58. [Google Scholar] [CrossRef]
  22. Markowitz, Harry M. 1956. The optimization of a quadratic function subject to linear constraints. Naval Research Logistics Quarterly 3: 111–33. [Google Scholar] [CrossRef]
  23. Maturo, Fabrizio, and Pierpaolo Angelini. 2023. Aggregate bound choices about random and nonrandom goods studied via a nonlinear analysis. Mathematics 11: 2498. [Google Scholar] [CrossRef]
  24. McCullagh, Peter. 1983. Quasi-likelihood functions. The Annals of Statistics 11: 59–67. [Google Scholar] [CrossRef]
  25. Nelder, John A., and Daryl Pregibon. 1987. An extended quasi-likelihood function. Biometrika 74: 221–32. [Google Scholar] [CrossRef]
  26. Nelder, John A., and Robert W. M. Wedderburn. 1972. Generalized linear models. Journal of the Royal Statistical Society, Series A (General) 135: 370–84. [Google Scholar] [CrossRef]
  27. Puntanen, Simo, and George P. H. Styan. 1989. The equality of the ordinary least squares estimator and the best linear unbiased estimator. The American Statistician 43: 153–61. [Google Scholar] [CrossRef]
  28. von Rosen, Dietrich. 1989. Maximum likelihood estimators in multivariate linear normal models. Journal of Multivariate Analysis 31: 187–200. [Google Scholar] [CrossRef]
  29. Watson, Geoffrey S. 1967. Linear least squares regression. The Annals of Mathematical Statistics 38: 1679–99. [Google Scholar] [CrossRef]
  30. Yuan, Ming, Ali Ekici, Zhaosong Lu, and Renato Monteiro. 2007. Dimension reduction and coefficient estimation in multivariate linear regression. Journal of the Royal Statistical Society, Series B (Statistical Methodology) 69: 329–46. [Google Scholar] [CrossRef]
  31. Yuan, Ming, and Yi Lin. 2006. Model selection and estimation in regression with grouped variables. Journal of the Royal Statistical Society, Series B (Statistical Methodology) 68: 49–67. [Google Scholar] [CrossRef]
  32. Yule, G. Udny. 1897. On the theory of correlation. Journal of the Royal Statistical Society 60: 812–54. [Google Scholar] [CrossRef]
Table 1. How to obtain   1 X using the notion of α -product.
Table 1. How to obtain   1 X using the notion of α -product.
Vector 211Sum
Vector 1
  1 X 1 101
  1 X n 011
Sum11n
Table 2. How to obtain   1 X   2 X using the notion of α -product.
Table 2. How to obtain   1 X   2 X using the notion of α -product.
Vector 2   2 X 1   2 X n Sum
Vector 1
  1 X 1 101
  1 X n 011
Sum11n
Table 3. How two marginal previsions are on y = x .
Table 3. How two marginal previsions are on y = x .
Vector 20234Sum
Vector 1
000000
20 0.2 00 0.2
300 0.3 0 0.3
4000 0.5 0.5
Sum0 0.2 0.3 0.5 1
Table 4. How two marginal previsions are on y = 2 x .
Table 4. How two marginal previsions are on y = 2 x .
Vector 20246Sum
Vector 1
000000
10 0.05 00 0.05
200 0.3 0 0.3
3000 0.65 0.65
Sum0 0.05 0.3 0.65 1
Table 5. The mean values of   1 X and   2 X do not depend on the joint probabilities.
Table 5. The mean values of   1 X and   2 X do not depend on the joint probabilities.
Vector 20224163Sum
Vector 1
0 p 11 p 12 p 13 p 14 0
10 p 21 p 22 p 23 p 24 0.2
20 p 31 p 32 p 33 p 34 0.3
30 p 41 p 42 p 43 p 44 0.5
Sum0 0.1 0.1 0.8 1
Table 6. The mean values of   1 X and   3 X do not depend on the joint probabilities.
Table 6. The mean values of   1 X and   3 X do not depend on the joint probabilities.
Vector 30264568Sum
Vector 1
0 p 11 p 12 p 13 p 14 0
10 p 21 p 22 p 23 p 24 0.2
20 p 31 p 32 p 33 p 34 0.3
30 p 41 p 42 p 43 p 44 0.5
Sum0 0.3 0.3 0.4 1
Table 7. The mean values of   2 X and   3 X do not depend on the joint probabilities.
Table 7. The mean values of   2 X and   3 X do not depend on the joint probabilities.
Vector 30264568Sum
Vector 2
0 p 11 p 12 p 13 p 14 0
22 p 21 p 22 p 23 p 24 0.1
41 p 31 p 32 p 33 p 34 0.1
63 p 41 p 42 p 43 p 44 0.8
Sum0 0.3 0.3 0.4 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Angelini, P. Extended Least Squares Making Evident Nonlinear Relationships between Variables: Portfolios of Financial Assets. J. Risk Financial Manag. 2024, 17, 336. https://doi.org/10.3390/jrfm17080336

AMA Style

Angelini P. Extended Least Squares Making Evident Nonlinear Relationships between Variables: Portfolios of Financial Assets. Journal of Risk and Financial Management. 2024; 17(8):336. https://doi.org/10.3390/jrfm17080336

Chicago/Turabian Style

Angelini, Pierpaolo. 2024. "Extended Least Squares Making Evident Nonlinear Relationships between Variables: Portfolios of Financial Assets" Journal of Risk and Financial Management 17, no. 8: 336. https://doi.org/10.3390/jrfm17080336

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop