Next Article in Journal
On the Calculation of Solid-Fluid Contact Angles from Molecular Dynamics
Next Article in Special Issue
Information Bottleneck Approach to Predictive Inference
Previous Article in Journal
Correlation Distance and Bounds for Mutual Information
Previous Article in Special Issue
Quantifying Morphological Computation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Consideration on Singularities in Learning Theory and the Learning Coefficient

Department of Mathematics, College of Science & Technology, Nihon University, 1-8-14, Surugadai, Kanda, Chiyoda-ku, Tokyo 101-8308, Japan
Entropy 2013, 15(9), 3714-3733; https://doi.org/10.3390/e15093714
Submission received: 21 June 2013 / Revised: 29 August 2013 / Accepted: 30 August 2013 / Published: 6 September 2013
(This article belongs to the Special Issue The Information Bottleneck Method)

Abstract

:
We consider the learning coefficients in learning theory and give two new methods for obtaining these coefficients in a homogeneous case: a method for finding a deepest singular point and a method to add variables. In application to Vandermonde matrix-type singularities, we show that these methods are effective. The learning coefficient of the generalization error in Bayesian estimation serves to measure the learning efficiency in singular learning models. Mathematically, the learning coefficient corresponds to a real log canonical threshold of singularities for the Kullback functions (relative entropy) in learning theory.

1. Introduction

The purpose of a learning system is to estimate an unknown true density function (a probability model) that generates the data. Real data associated with, for example, genetic analysis, data mining, image or speech recognition, artificial intelligence, the control of a robot and time series prediction are very complicated and usually are not generated by a simple normal distribution. In Bayesian estimation, we set a learning model that is written in probabilistic form with parameters, and our goal is to estimate the true density function by a predictive function constructed with the learning model and such data. Therefore, the learning model should be abundant enough to capture the true density function’s structure. Hierarchical learning models, such as the layered neural network, the Boltzmann machine, the reduced rank regression and the normal mixture model, are known to be effective learning models for analyzing such data. These are, however, singular learning models, which cannot be analyzed using the classic theory of regular statistical models, because singular learning models have a singular Fisher metric that is not always approximated by any quadratic form [1,2,3,4]. Therefore, it is difficult to analyze their generalization errors, which indicate how precisely the predictive function approximates the true density function.
In recent studies, Watanabe showed using algebraic geometry that the generalization and training errors are subject to a universal law and defined the model selection method “widely applicable information criterion” (WAIC ) as a generalized Akaike information criterion (AIC) [5,6,7,8,9]. WAIC can even be applied to singular learning models, whereas AIC cannot. Using the WAIC, we can estimate the generalization errors from the training errors without any knowledge of the true probability density functions. The generalization errors relate to the generalization losses via the entropy of the true distribution. Thus, we can select a suitable model from among several statistical models by this method.
Computing the WAIC requires the values of the learning coefficient and the singular fluctuation, which are both birational invariants. Mathematically, the learning coefficient is the log canonical threshold (Definition 1) of the Kullback function (relative entropy), and the singular fluctuation is known as a statistically generalized log canonical threshold, which is obtained theoretically from the learning coefficient (Equation (1) in Section 2). These values can be obtained by Hironaka’s Theorem (Appendix A). However, it is still difficult to obtain these within learning theory for several reasons, such as degeneration with respect to their Newton polyhedra and non-isolation of their singularities [10]. Moreover, in algebraic geometry and algebraic analysis, these studies are usually done over an algebraically closed field [11,12]; many differences exist for real and complex fields. For example, log canonical thresholds over the complex field are less than one, whereas those over the real field are not necessarily so. We, therefore, cannot apply results over an algebraically closed field to our current situation directly (Appendix B). One of the bottlenecks in learning theory is to obtain the learning coefficients and the singular fluctuation.
In this paper, we consider the learning coefficient of “Vandermonde matrices-type singularities” in statistical learning theory. The reason why we contribute only to such singularities is that the Vandermonde matrix type is generic and essential in learning theory. These log canonical thresholds give the learning coefficients of normal mixture models, three-layered neural networks and mixtures of binomial distributions, which are widely used as effective learning models (Section 3.1 and Section 3.2 and [13]). Moreover, we prove Theorem 2 (the method for finding a deepest deepest singular point) and Theorem 3 (the method to add variables), which are very beneficial to obtain the log canonical threshold for the homogeneous case. Theorem 2 indicates the best point of singularities that gives the log canonical threshold. Therefore, this theorem is useful for the reduction of the number of blowup processes. Theorem 3 improves our recursive blowup method by simplifying coordinate system changes with added variables. These two theorems enable us to obtain a new bound for the log canonical thresholds of Vandermonde matrix-type singularities in Theorem 5. These bounds are much tighter than those in [14].
In the past few years, we have obtained the learning coefficients for reduced rank regression [15], for the three-layered neural network with one input unit and one output unit [16,17], and for the normal mixture models with a dimension of one [18]. The paper [14] derived bounds on the learning coefficients for the Vandermonde matrix-type singularities and explicit values under some conditions. The learning coefficients for the restricted Boltzmann machine [19] have also been considered recently. Ref [20,21,22], respectively, obtained these for naive Bayesian networks and for directed tree models with hidden variables. These results give partial answers for the learning coefficients.
The rest of the paper is in three sections. Section 2 summarizes the framework of Bayesian learning models. In Section 3, we demonstrate our main theorems and consider the log canonical threshold of Vandermonde matrix-type singularities (Definition 3). We finish with our conclusions in Section 4.

2. Learning Coefficients and Singular Fluctuations

In this section, we present the theory of learning coefficients and singular fluctuations. Let q ( x ) be a true probability density function of variables, x R N , and let x n : = { x i } i = 1 n be n training samples selected from q ( x ) independently and identically. Consider a learning model that is written in probabilistic form as p ( x | w ) , where w W R d is a parameter. The purpose of the learning system is to estimate q ( x ) from x n using p ( x | w ) . Let ψ ( w ) be an a priori probability density function on the parameter set, W, and p ( w | x n ) be the a posteriori probability density function:
p ( w | x n ) = 1 Z n ψ ( w ) i = 1 n p ( x i | w )
where:
Z n = W ψ ( w ) i = 1 n p ( x i | w ) d w
Let us define for the inverse temperature, β:
E w [ f ( w ) ] = d w f ( x ) ψ ( w ) i = 1 n p ( x i | w ) β d w ψ ( w ) i = 1 n p ( x i | w ) β
We usually set β = 1 .
We then have a predictive density function, p ( x | X n ) = E w [ p ( x | w ) ] , which is the average inference of the Bayesian density function.
We next introduce the Kullback function, K ( q | | p ) , and the empirical Kullback function, K n ( q | | p ) , for density functions p ( x ) , q ( x ) :
K ( q | | p ) = x q ( x ) log q ( x ) p ( x )
K n ( q | | p ) = 1 n i = 1 n log q ( x i ) p ( x i )
The function, K ( p | | q ) , always has a non-negative value and satisfies K ( q | | p ) = 0 , if and only if q ( x ) = p ( x ) .
The Bayesian generalization error, B g , Bayesian training error, B t , Gibbs generalization error, G g , and Gibbs training error, G t , are defined as follows:
B g = K ( q ( x ) | | E w [ p ( x | w ) ] ) )
B t = K n ( q ( x ) | E w [ p ( x i | w ) ] )
G g = E w [ K ( q ( x ) | | p ( x | w ) ) ]
and
G t = E w [ K n ( q ( x ) | | p ( x | w ) ) ]
The most important of these is the Bayesian generalization error. This error describes how precisely the predictive function approximates the true density function.
Watanabe [6,7,23] proved the following four relations:
E [ B g ] = λ + ν β ν n β + o ( 1 n )
E [ B t ] = λ ν β ν n β + o ( 1 n )
E [ G g ] = λ + ν β n β + o ( 1 n )
E [ G t ] = λ ν β n β + o ( 1 n )
Thus we have:
E [ B g ] = E [ B t ] + 2 β ( E [ G t ] E [ B t ] ) + o ( 1 n )
and
E [ G g ] = E [ G t ] + 2 β ( E [ G t ] E [ B t ] ) + o ( 1 n )
Eliminating the expectation of the true probability density function from the above four errors and setting:
B L g = x q ( x ) log E w [ p ( x | w ) ]
B L t = 1 n i = 1 n log E w [ p ( x i | w ) ]
G L g = E w [ x q ( x ) log p ( x | w ) ]
G L t = E w [ 1 n i = 1 n log p ( x i | w ) ]
we then have:
E [ B L g ] = E [ B L t ] + 2 β ( E [ G t ] E [ B t ] ) + o ( 1 n )
and
E [ G L g ] = E [ G L t ] + 2 β ( E [ G t ] E [ B t ] ) + o ( 1 n )
These two equations constitute the WAIC and show that we can estimate the Bayesian and Gibbs generalization errors from the Bayesian and Gibbs training errors without any knowledge of the true probability density functions. Training errors are calculated from training samples, x i , using a learning model, p. In real applications or experiments, we usually do not know the true distribution, but only the values of the training errors. Our purpose is to estimate the true distribution from the training samples, showing that these relations are effective. We can select a suitable model from among several statistical models by observing these values.
Let λ denote a learning coefficient and ν a singular fluctuation, both of which are birational invariants. Mathematically, λ is equal to the log canonical threshold introduced in Definition 1 and Appendix B. For regular models, λ = ν = d / 2 holds, where d is the dimension of the parameter space.
The difference between the Bayesian and Gibbs training errors converges to ν / n :
n β ( E [ G t ] E [ B t ] ) ν
These relations were shown using the resolution of singularities and the Schwarz distribution.
From the learning coefficient, λ, and its order, θ, the value, ν, is obtained theoretically as follows. Let ξ ( u ) be an empirical process defined on the manifold obtained by a resolution of singularities, and u * denote the sum of local coordinates that attain the minimum λ and the maximum θ. We then have:
ν = 1 2 E ξ 0 d t u * d u ξ ( u ) t λ 1 / 2 e β t + β t ξ ( u ) 0 d t u * d u t λ 1 / 2 e β t + β t ξ ( u )
ξ ( u ) is a random variable of a Gaussian process with mean zero and variance two. Our purpose in this paper is to obtain λ.
To assist in achieving this aim, we use the desingularization approach from algebraic geometry (cf. Appendix A). It is a new problem in algebraic geometry to obtain the desingularization of the Kullback functions, because the singularities of these functions are very complicated, and as such, most of these have not yet been investigated.

3. Main Theorems and Vandermonde Matrix-Type Singularities

We denote constants, such as a * , b * and w * , by the suffix ∗. Additionally, for simplicity, we use the notation: w = { a k i , b i j } 1 i H instead of: w = { a k i , b i j } 1 k M , 1 i H , 1 j N , because we always have 1 k M and 1 j N in this paper.
Define the norm of a matrix, C = ( c i j ) , by | | C | | = i , j | c i j | 2 . Set N + 0 = N { 0 } .
Definition 1 
For a real analytic function, f, in a neighborhood, U, of w * and a C function ψ with a compact support, let λ w * ( f , ψ ) be the largest pole of U | f | z ψ d w and θ w * ( f , ψ ) be its order. If ψ ( w * ) 0 , then we denote λ w * ( f ) = λ w * ( f , ψ ) and θ w * ( f ) = θ w * ( f , ψ ) , because the log canonical threshold and its order are independent of ψ.
Definition 2 
Fix Q N . Define: [ b 1 * , b 2 * , , b N * ] Q = γ i ( 0 , , 0 , b i * , , b N * ) if b 1 * = = b i 1 * = 0 , b i * 0 , and γ i = 1 i f   Q   i s   o d d , | b i * | / b i * i f   Q   i s   e v e n .
Definition 3 
Fix Q N .
Let A = a 11 a 1 H a 1 , H + 1 * a 1 , H + r * a 21 a 2 H a 2 , H + 1 * a 2 , H + r * a M 1 a M H a M , H + 1 * a M , H + r * , I = ( 1 , , N ) N + 0 N
B I = ( j = 1 N b 1 j j , j = 1 N b 2 j j , , j = 1 N b H j j , j = 1 N b H + 1 , j * j , , j = 1 N b H + r , j * j ) t
and
B = ( B I ) 1 + + N = Q n + 1 , 0 n H + r 1 = ( B ( 1 , 0 , , 0 ) , B ( 0 , 1 , , 0 ) , , B ( 0 , 0 , , 1 ) , B ( 1 + Q , 0 , , 0 ) , )
(the superscript, t, denotes matrix transposition).
a k i and b i j ( 1 k M , 1 i H , 1 j N ) are variables in a neighborhood of a k i * and b i j * , where a k i * and b i j * are fixed constants.
Let I be the ideal generated by the elements of A B .
We call singularities of I Vandermonde matrix-type singularities.
To simplify, we usually assume that
( a 1 , H + j * , a 2 , H + j * , , a M , H + j * ) t 0 , ( b H + j , 1 * , b H + j , 2 * , , b H + j , N * ) 0
for 1 j r and
[ b H + j , 1 * , b H + j , 2 * , , b H + j , N * ] Q [ b H + j , 1 * , b H + j , 2 * , , b H + j , N * ] Q
for j j .
Example 1 
If N = Q = 1 and r = 0 , then we have: B = b 11 b 11 2 b 11 H b 21 b 21 2 b 21 H b H 1 b H 1 2 b H 1 H .
This matrix is a Vandermonde matrix.
Example 2 
If Q = 1 , M = 1 , H = 2 , N = 2 and r = 1 , then we have: A = a 11 a 12 a 1 , 3 * and B = b 11 b 12 b 11 2 b 11 b 12 b 12 2 b 11 3 b 11 b 12 2 b 11 2 b 12 b 12 3 b 21 b 22 b 21 2 b 21 b 22 b 22 2 b 21 3 b 21 b 22 2 b 21 2 b 22 b 22 3 b 31 * b 32 * b 31 * 2 b 31 * b 32 * b 32 * 2 b 31 * 3 b 31 * b 32 * 2 b 31 * 2 b 32 * b 32 * 3 .
In this paper, we denote:
A M , H = a 11 a 12 a 1 H a 21 a 22 a 2 H a M 1 a M 2 a M H , B H , N , I = j = 1 N b 1 j j j = 1 N b 2 j j j = 1 N b H j j a n d
B H , N ( Q ) = ( B H , N , I ) 1 + + N = Q n + 1 , 0 n H 1 .
Furthermore, we denote: a * = a 1 , H + 1 * a M , H + 1 * and
( A M , H , a * ) = a 11 a 12 a 1 H a 1 , H + 1 * a 21 a 22 a 2 H a 2 , H + 1 * a M 1 a M 2 a M H a M , H + 1 *
Theorem 1 
([18]) Consider a sufficiently small neighborhood, U, of
w * = { a k i * , b i j * } 1 i H
and variables, w = { a k i , b i j } 1 i H , in the set, U.
Set: ( b 01 * * , b 02 * * , , b 0 N * * ) = ( 0 , , 0 ) .
Let each: ( b 11 * * , b 12 * * , , b 1 N * * ) , …, ( b r 1 * * , b r 2 * * , , b r N * * ) be a different real vector in:
[ b i 1 * , b i 2 * , , b i N * ] Q 0 , f o r   i = 1 , , H + r
That is:
{ ( b 11 * * , , b 1 N * * ) , , ( b r 1 * * , , b r N * * ) ; [ b i 1 * , , b i N * ] Q 0 , i = 1 , , H + r }
Then, r is uniquely determined, and r r by the assumption in Definition 3. Set: ( b i 1 * * , , b i N * * ) = [ b H + i , 1 * , , b H + i , N * ] Q , for 1 i r .
Assume that:
[ b i 1 * , , b i N * ] Q = 0 , 1 i H 0 ( b 11 * * , , b 1 N * * ) , H 0 + 1 i H 0 + H 1 , ( b 21 * * , , b 2 N * * ) , H 0 + H 1 + 1 i H 0 + H 1 + H 2 , ( b r 1 * * , , b r N * * ) , H 0 + + H r 1 + 1 i H 0 + + H r ,
and H 0 + + H r = H .
We then have:
λ w * ( | | A B | | 2 ) = M r 2 + λ w 1 ( 0 ) * ( | | A M , H 0 B H 0 , N ( Q ) | | 2 ) + α = 1 r λ w 1 ( α ) * ( | | ( A M , H α 1 , a ( α ) * ) B H α , N ( 1 ) | | 2 ) + α = r + 1 r λ w 1 ( α ) * ( | | A M , H α 1 B H α 1 , N ( 1 ) | | 2 )
where: w 1 ( 0 ) * = { a k , i * , 0 } 1 i H α , w 1 ( α ) * = { a k , H 0 + + H α 1 + i * , 0 } 2 i H α and a ( α ) * = a 1 , H + α * a M , H + α * for α 1 .
Theorem 2 (Method for finding a deepest singular point) 
Let f 1 ( w 1 , , w d ) , …, f m ( w 1 , , w d ) be homogeneous functions of w 1 , , w j ( j d ) with the degree, n i , of w 1 , , w j . Furthermore, let ψ be a C function, such that ψ ( 0 , , 0 , w j + 1 * , , w d * ) ψ ( w 1 * , , w d * ) and ψ w is homogeneous of w 1 , , w j in a small neighborhood of ( 0 , , 0 , w j + 1 * , , w d * ) .
Then, we have:
λ ( 0 , , 0 , w j + 1 * , , w d * ) ( f 1 2 + + f m 2 , ψ ) λ ( w 1 * , , w j * , w j + 1 * , , w d * ) ( f 1 2 + + f m 2 , ψ )
(Proof)
Let d be the degree of w 1 , , w j for ψ in a neighborhood of ( 0 , , 0 , w j + 1 * , , w d * ) . Let us construct the blowup of f 1 , , f m along the submanifold, { v = 0 , w i = 0 , 1 i j } . Let w i = v w i for 1 i j . We have: v n i f i ( w 1 , , w d ) and ( f 1 ( w ) 2 + f 2 ( w ) 2 + + f m ( w ) 2 ) z ψ d w d v = ( v 2 n 1 f 1 2 ( w ) + v 2 n 2 f 2 2 ( w ) + + v 2 n m f m ( w ) 2 ) z ψ ( w ) v d + j d w d v . Because: v 2 n i f i 2 ( w 1 , , w d ) f i 2 ( w 1 , , w d ) for | v | < 1 , we have: v 2 n 1 f 1 2 + + v 2 n m f m 2 f 1 2 + + f m 2 , and, hence, by Lemma 1 in Appendix C:
λ ( 0 , , 0 , w j + 1 * , , w d * ) ( f 1 2 + + f m 2 , ψ ) min { d + j + 1 , λ w * ( f 1 2 + + f m 2 , ψ ) }
Furthermore, we consider the construction of the blowup of: f 1 , , f m along the submanifold: { w i = 0 , 1 i j } , for which we have
λ ( 0 , , 0 , w j + 1 * , , w d * ) ( f 1 2 + + f m 2 , ψ ) d + j       Q.E.D.
In general, it is not true that:
λ w 0 ( f 1 2 + + f m 2 , ψ ) λ w * ( f 1 2 + + f m 2 , ψ )
even if w 0 R d satisfies:
f i ( w 0 ) = f i w j ( w 0 ) = 0 , 1 i m , 1 j d
Example 3 
Let f 1 = x ( x 1 ) 2 , f 2 = ( y 2 + ( x 1 ) 2 ) ( ( y 1 ) 6 + x ) and f 3 = ( z 2 + ( x 1 ) 2 ) ( ( z 1 ) 6 + x ) . Then, we have: f 1 = f 2 = f 3 = f 1 x = f 2 y = f 2 x = f 3 z = f 3 x = 0 if and only if x = 1 , y = 0 , z = 0 .
In this case, we have λ ( 1 , 0 , 0 ) ( f 1 2 + f 2 2 + f 3 2 ) = 1 / 4 + 1 / 4 + 1 / 4 > λ ( 0 , 1 , 1 ) ( f 1 2 + f 2 2 + f 3 2 ) = 1 / 2 + 1 / 12 + 1 / 12 .
Theorem 3 (Method to add variables) 
Let f 1 ( w 1 , , w d ) , …, f m ( w 1 , , w d ) be homogeneous functions of w 1 , , w d of the degree, n i , in w 1 , , w d . Set: f 1 ( w 2 , , w d ) = f 1 ( 1 , w 2 , , w d ) , …, f m ( w 2 , , w d ) = f m ( 1 , w 2 , , w d ) . If w 1 * 0 , then we have:
λ ( w 1 * , , w d * ) ( f 1 2 + + f m 2 ) = λ ( w 2 * / w 1 * , , w d * / w 1 * ) ( f 1 2 + + f m 2 )
(Proof) Set w 2 = w 2 / w 1 , , w d = w d / w 1 . Then, we have:
f i ( w 1 , w 2 , , w d ) = w 1 n i f i ( 1 , w 2 , , w d ) = w 1 n i f i
Since w 1 n i 0 on a small neighborhood of w 1 * , there exist positive real numbers, C , C , such that:
C ( f 1 2 + + f m 2 ) f 1 2 + + f m 2 C ( f 1 2 + + f m 2 )
This completes the proof by Lemma 1 in Appendix C. Q.E.D.
Remark 1
The above theorem shows that we can set nonzero constants as variables to obtain the same log canonical threshold. However, in general, this is not true.
(1)
Consider the function f ( x , y ) = ( y 1 + x 2 ) 2 . We have λ f = 0 ( f ) = 1 / 2 , whereas λ f ( x , 1 ) = 0 ( f ( x , 1 ) ) = 1 / 4 .
(2)
Consider the function f ( x 1 , x 2 , x 3 , x 4 , x 5 , y ) = ( x 1 2 + x 2 2 + x 3 2 + x 4 2 + x 5 2 + y 1 ) 2 . We have λ f = 0 ( f ) = 1 / 2 , whereas λ 0 ( f ( x 1 , x 2 , x 3 , x 4 , x 5 , 1 ) ) = 5 / 4 .
The second example shows that the following theorem over the complex field is not true over the real field.
Theorem 4 [11]
Let f ( x 1 , , x d , y ) be a holomorphic function near zero, and for a hyperplane H, let g = f | y = 0 (or g = f H ) denote the restriction of f to y = 0 (or H). Then, λ g = 0 ( g ) λ f = 0 ( f ) .
Define: k l = k ! l ! ( k l ) ! ,
Theorem 5 
We use the same notation as in Theorem 1. Let:
bound 1 = min { ( H i + 1 ) N + d i ( s ) + d i ( s ) + d i ( s ) 2 ( c o u n t ( i , s , k ( s ) ) 1 ) Q + 2 : 1 i s , 1 k ( 1 ) , , k ( s ) N , 1 s H }
where: c o u n t ( i , s , j ) = # { i 1 : i i 1 s , k ( i 1 ) = j } , C ( i , s ) = # { c o u n t ( i , s , j ) = 0 , 1 j N } ,
d i ( s ) = ( N 1 ) Q s 1 = i s ( c o u n t ( i , s 1 , k ( s 1 ) ) 1 ) d i ( s ) = M ( i 1 ) { ( c o u n t ( i , s , k ( s ) ) 1 ) Q + 1 } + Q M s 1 = i , c o u n t ( i , s , k ( s ) ) > c o u n t ( i , s 1 , k ( s 1 ) ) s 1 ( c o u n t ( i , s , k ( s ) ) c o u n t ( i , s 1 , k ( s 1 ) ) ) , d i ( s ) = 0 , i f   c o u n t ( i , s , k ( s ) ) = 1 , ( H s ) { C ( i , s ) Q + ( N 1 ) Q ( c o u n t ( i , s , k ( s ) ) 2 ) } , i f   c o u n t ( i , s , k ( s ) ) 2 , N 1 M , ( H s ) { C ( i , s ) Q + M Q ( c o u n t ( i , s , k ( s ) ) 2 ) } , i f   c o u n t ( i , s , k ( s ) ) 2 , C ( i , s ) M < N 1 , ( H s ) { M Q ( c o u n t ( i , s , k ( s ) ) 1 ) } , i f   c o u n t ( i , s , k ( s ) ) 2 , M C ( i , s ) .
Furthermore, let: bound 2 = N H + i = 0 k 1 M Q ( k i ) N + Q i N 1 2 + 2 Q k , where: k = max { i Z ; N H M i = 0 i 1 ( 1 + Q i ) N + Q i N 1 } , and let:
bound 3 = M H 2
We have
λ 0 ( | | A M , H B H , N ( Q , m ) | | 2 ) min { bound 1 , bound 2 , bound 3 } λ 0 ( | | ( A M , H 1 , a * ) B H , N ( Q , m ) | | 2 ) min { bound 1 , bound 2 }
The proof appears in Appendix C.
Figure 1a–d show the values of new bounds, min { bound 1 , bound 2 , bound 3 } , for (a) H = 7 , N = 6 , (b) H = 8 , N = 6 , (c) H = 7 , N = 7 and (d) H = 8 , N = 7 with Q = 2 , respectively. We compare these values with those obtained by the past work in [14]. In the figures, the horizontal axis is the number, M, and the vertical one, the value of such bounds. The dashed lines indicate the bounds obtained by the past work. These figures show that new bounds are not greater than old ones.
Figure 1. The values of new bounds, min { bound 1 , bound 2 , bound 3 } , for (a) H = 7 , N = 6 ; (b) H = 8 , N = 6 ; (c) H = 7 , N = 7 and (d) H = 8 , N = 7 with Q = 2 , compared with the bounds obtained by the past work in [14].
Figure 1. The values of new bounds, min { bound 1 , bound 2 , bound 3 } , for (a) H = 7 , N = 6 ; (b) H = 8 , N = 6 ; (c) H = 7 , N = 7 and (d) H = 8 , N = 7 with Q = 2 , compared with the bounds obtained by the past work in [14].
Entropy 15 03714 g001
In paper [24], we had exact values for N = 1 :
λ 0 ( | | A M , H B H , 1 ( Q ) | | 2 ) = M Q k ( k + 1 ) + 2 H 4 ( 1 + k Q )
where: k = max { i Z : 2 H M ( i ( i 1 ) Q + 2 i ) } , and we had:
θ = 1 , if  2 H > M ( k ( k 1 ) Q + 2 k ) 2 , if  2 H = M ( k ( k 1 ) Q + 2 k )
We had other exact values when H is small on paper [14]. Both sets of exact values are the bounded values in Theorem 5.

3.1. A Learning Coefficient for a Three-Layered Neural Network

Consider the three-layered neural network with N input units, H hidden units and M output units, which are trained for estimating the true distribution with r hidden units. Their learning coefficients, λ, are as follows [14,24]:
λ = min { λ 0 ( | | A M , H 0 B H 0 , N ( 2 ) | | 2 ) + α = 1 r λ 0 ( | | ( A M , H α 1 , a * ) B H α , N ( 1 ) | | 2 ) : H 0 + H 1 + + H r = H ,   a * 0 } .

3.2. A Learning Coefficient for a Normal Mixture Model

Consider normal mixture models with H peaks and the true distribution with r peaks. Then, their learning coefficients, λ, are as follows [14,18]:
λ = r 1 2 + min { α = 1 r λ 0 ( | | ( A M , H α 1 , a * ) B H α , N ( 1 ) | | 2 ) : α = 1 r H α = H , a * 0 }
In particular, we have for N = 1 :
λ = r 1 + i + i 2 + 2 ( H ( r 1 ) ) 4 ( i + 1 ) , θ = 1 , if  i 2 + i < 2 ( H ( r 1 ) ) , or  H = r 2 , if  i 2 + i = 2 ( H ( r 1 ) ) ,
where i = max { j Z ; j 2 + j 2 ( H ( r 1 ) ) } .

4. Conclusions

In this paper, we prove two theorems, Theorem 2 (the method for finding a deepest singular point) and Theorem 3 (the method to add variables) for obtaining learning coefficients in a homogeneous case. By applying these methods to Vandermonde matrix-type singularities and using the inclusion of ideals and recursive blowup from algebraic geometry, we found new bounds on learning coefficients for Vandermonde matrix-type singularities. These bounds are much tighter than those in [14]. Our future research aim is to improve our methods and to obtain exact values for the general machine model.
The learning coefficients from our recent results have been used very effectively by Drton [25,26] for model selection, using a method called “singular Bayesian information criterion (sBIC)”, which can be applied to singular models, where the assumptions supporting the use of the standard BIC do not hold. Our theoretical results introduce a mathematical measure of precision to numerical calculations, such as Markov chain Monte Carlo (MCMC). Nagata and Watanabe [27,28] gave a mathematical foundation for analyzing and developing the precision of the MCMC method using our theoretical values of marginal likelihoods.

Acknowledgments

This research was supported by the Ministry of Education, Culture, Sports, Science and Technology in Japan, Grant-in-Aid for Scientific Research 22540224.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

We introduce Hironaka’s Theorem on desingularization.
Theorem 6 
[Desingularization, Hironaka (1964), (Figure A1)]
Let f be a real analytic function in a neighborhood of w * = ( w 1 * , , w d * ) R d with f ( w * ) = 0 . There exists an open set, V w * , a real analytic manifold, U, and a proper analytic map, μ, from U to V, such that:
(1) 
μ : U E V f 1 ( 0 ) is an isomorphism, where E = μ 1 ( f 1 ( 0 ) ) ,
(2) 
for each u U , there is a local analytic coordinate system ( u 1 , , u n ) , such that f ( μ ( u ) ) = ± u 1 s 1 u 2 s 2 u n s n , where s 1 , , s n are non-negative integers.
Figure A1. Hironaka’s Theorem: diagram of desingularization, μ, of f: E maps to f 1 ( 0 ) . U E is isomorphic to V f 1 ( 0 ) by μ, where V is a small neighborhood of w * with f ( w * ) = 0 .
Figure A1. Hironaka’s Theorem: diagram of desingularization, μ, of f: E maps to f 1 ( 0 ) . U E is isomorphic to V f 1 ( 0 ) by μ, where V is a small neighborhood of w * with f ( w * ) = 0 .
Entropy 15 03714 g002

Appendix B

The learning coefficient is the log canonical threshold of the Kullback function (relative entropy). In this section, we explain its difference for real and complex fields. Let f be a nonzero holomorphic function over C or an analytic function over R on a smooth variety, Y, and let Z Y be a closed subscheme. The log canonical threshold, λ Z ( Y , f ) , is defined analytically as:
λ Z ( Y , f ) = sup { c : | f | c  is locally  L 2  near  Z }
over C , and:
λ Z ( Y , f ) = sup { c : | f | c  is locally  L 1  near  Z }
over R [11,12]. It is known that if f is a polynomial or a convergent power series, then λ 0 ( C d , f ) is the largest root of the Bernstein-Sato polynomial, b ( s ) C [ s ] , of f, where b ( s ) f s = P f s + 1 for a linear differential operator, P [29,30,31]. The log canonical threshold, λ Z ( Y , f ) , also corresponds to the largest pole of near Z | f | 2 z ψ ( w ) d w over C , ( near Z | f | z ψ ( w ) d w over R ), where ψ ( w ) is a C function with a compact support, such that ψ ( w ) 0 on Z.

Appendix C

Using the blowup process and the method to add variables together with the inductive method for s, we demonstrate Theorem 5
We give below Lemma 1, as it is frequently used in the proofs.
Lemma 1 
([18,24,32]) Let U be a neighborhood of w * R d . Let I be the ideal generated by f 1 , , f n , which are analytic functions defined on U.
( 1 ) If g 1 2 + + g m 2 f 1 2 + + f n 2 , then λ w * ( g 1 2 + + g m 2 ) λ w * ( f 1 2 + + f n 2 ) .
( 2 ) If g 1 , , g m I , then λ w * ( g 1 2 + + g m 2 ) λ w * ( f 1 2 + + f n 2 ) . In particular, if g 1 , , g m generate the ideal I, then λ w * ( f 1 2 + + f n 2 ) = λ w * ( g 1 2 + + g m 2 ) .
The following lemma is also used in the proofs.
Lemma 2 
([19]) Let I , J be the ideals generated by f 1 ( w ) , , f n ( w ) and g 1 ( w ) , , g m ( w ) , respectively. If w and w are different variables, then
λ ( w * , w * ) ( f 1 2 + + f n 2 + g 1 2 + + g m 2 ) = λ w * ( f 1 2 + + f n 2 ) + λ w * ( g 1 2 + + g m 2 ) .
Step 1
Let us consider the following procedure from s = 1 to s = H , and the generators of the ideal:
J = a i 0 1 a i 0 H b 11 1 b 1 N N b 21 1 b 2 N N b H 1 1 b H N N : 1 i 0 M , i = 1 N i = n Q + 1 , n 0 .
By constructing the blowup repeatedly and choosing one branch of the blowup process, we show the following (i)∼(v) in this subsection:
(i)
k ( 1 ) , , k ( s 1 ) { 1 , 2 , , N } ,
(ii)
c o u n t ( i 1 , i 2 , j ) = # { k ( i ) = j | i 1 i i 2 } for 1 j N ,
(iii)
b i j = v 1 v 2 v i b i j , i < s 1 , v 1 v 2 v s 1 b i j , i s 1 , and b i k ( i ) = 1 ,
(iv)
The Jacobian is:
i = 1 H j = 1 N d b i j = i = 1 s 1 v i ( H i + 1 ) N 1 + d i d v i i = 1 s 1 j = 1 N d b i j
and
d i = ( N 1 ) Q s 1 = i s 1 ( c o u n t ( i , s 1 , k ( s 1 ) ) 1 ) ,
(v)
J = a i 0 i 1 b i 1 k ( i 1 ) k ( i ) = k ( i 1 ) , 1 i i 1 1 ( b i 1 , k ( i 1 ) Q b i , k ( i 1 ) Q ) : 1 i 0 M , 1 i 1 s 1 + a i 0 s a i 0 H b s j n Q + 1 k ( i ) = j , 1 i s 1 ( b s j Q b i j Q ) b H j n Q + 1 k ( i ) = j , 1 i s 1 ( b H j Q b i j Q ) : 1 i 0 M , j = 1 , , N , n 0 + a i 0 s a i 0 H b s 1 1 b s N N b H 1 1 b H N N : 1 i 0 M , i = 1 N i = n Q + 1 , n 1 , j < Q n + 1 .
By Theorem 3, we can set b i k ( i ) as a variable.
Now, we show the above by the inductive method.
Define 1 k ( s ) N . Construct the blowup along { b i j = 0 , s i H , 1 j N } . Set b i j = v s b i j for s i H , 1 j N , and set b s k ( s ) = 1 .
By constructing the blowup along { v i = b s j = 0 , 1 j N , j k ( s ) } repeatedly, and by choosing one branch of the blowup process, set b s j = b s j i = 1 s 1 v i d i for j k ( s ) , where d i = ( c o u n t ( i , s , k ( s ) ) 1 ) Q .
Consider a sufficiently small neighborhood of { v i = 0 } using Theorem 2.
Set f i k ( s ) = b i k ( s ) k ( i 1 ) = k ( s ) , 1 i 1 s 1 ( b i k ( s ) Q b i 1 k ( s ) Q ) for i s , b ˜ i j = b i j b s j f i k ( s ) f s k ( s ) and b ˜ i j = b ˜ i j i 1 = 1 s v i 1 = b i j b s j f i k ( s ) f s k ( s ) for i s + 1 , j k ( s ) .
We then have:
b s 1 1 b s N N b H 1 1 b H N N = j = 1 N b s j j b s + 1 , k ( s ) k ( s ) j = 1 , j k ( s ) N ( b ˜ s + 1 , j + b s j f s + 1 , k ( s ) f s k ( s ) ) j b H , k ( s ) k ( s ) j = 1 N ( b ˜ H j + b s j f H k ( s ) f s k ( s ) ) j
which is an element of the vector ideal:
j = 1 N b s j j j = 1 , j k ( s ) N b s j j b s + 1 , k ( s ) k ( s ) ( f s + 1 , k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j j = 1 , j k ( s ) N b s j j b H k ( s ) k ( s ) ( f H k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j + k ( s ) = k ( s ) , 0 j j , j j 0 j = 1 , j k ( s ) N b s j j b s + 1 , k ( s ) k ( s ) j = 1 , j k ( s ) N b ˜ s + 1 , j j j ( f s + 1 , k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j j = 1 , j k ( s ) N b s j j b H k ( s ) k ( s ) j = 1 , j k ( s ) N b ˜ H j j j ( f s + 1 , k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j
Furthermore, we have: j = 1 N b s j j j = 1 , j k ( s ) N b s j j b s + 1 , k ( s ) k ( s ) ( f s + 1 , k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j j = 1 , j k ( s ) N b s j j b H k ( s ) k ( s ) ( f H k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j is an element of:
j = 1 , j k ( s ) N b s j j f s k ( s ) j = 1 , j k ( s ) N j b s , k ( s ) n Q + 1 k ( i 1 ) = k ( s ) , 1 i 1 s 1 ( b s k ( s ) Q b i 1 k ( s ) Q ) b s + 1 , k ( s ) n Q + 1 k ( i 1 ) = k ( s ) , 1 i 1 s 1 ( b s + 1 k ( s ) Q b i 1 k ( s ) Q ) b H k ( s ) n Q + 1 k ( i 1 ) = k ( s ) , 1 i 1 s 1 ( b H k ( s ) Q b i 1 k ( s ) Q ) : n 0
Since b s j = b s j i = 1 s 1 v i d i for j k ( s ) , where d i = ( c o u n t ( i , s , k ( s ) ) 1 ) Q , we have b s j = b s j i = 1 s 1 v i d i + 1 and j = 1 , j k ( s ) N b s j j f s k ( s ) j = 1 , j k ( s ) N j is finite. That is, we have:
j = 1 N b s j j j = 1 , j k ( s ) N b s j j b s + 1 , k ( s ) k ( s ) ( f s + 1 , k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j j = 1 , j k ( s ) N b s j j b H k ( s ) k ( s ) ( f H k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j b s , k ( s ) n Q f s k ( s ) b s + 1 , k ( s ) n Q f s + 1 , k ( s ) b H k ( s ) n Q f H k ( s ) : n 0
If we assume that for α:
J α = b s k ( s ) n Q f s k ( s ) b H k ( s ) n Q f H k ( s ) , n 0 + b s 1 1 b s N N b H 1 1 b H N N : j = 1 N j = n Q + 1 , n 1 , j = 1 , j k ( s ) N j α = b s k ( s ) n Q f s k ( s ) b H k ( s ) n Q f H k ( s ) , n 0 + 0 b s + 1 , k ( s ) j = 1 , j k ( s ) N b ˜ s + 1 , j j b H k ( s ) j = 1 , j k ( s ) N b ˜ H j j : j = 1 N j = n Q + 1 , n 1 , j = 1 , j k ( s ) N j α ,
we have for j = 1 , j k ( s ) N j = α + 1 :
b s 1 1 b s N N b H 1 1 b H N N + J α = 0 b s + 1 , k ( s ) j = 1 , j k ( s ) N b ˜ s + 1 , j j b H k ( s ) j = 1 , j k ( s ) N b ˜ H j j + J α
since for j 0 , we have:
0 j = 1 , j k ( s ) N b s j j b s + 1 , k ( s ) k ( s ) j = 1 , j k ( s ) N b ˜ s + 1 , j j j ( f s + 1 , k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j j = 1 , j k ( s ) N b s j j b H k ( s ) k ( s ) j = 1 , j k ( s ) N b ˜ H j j j ( f H , k ( s ) f s k ( s ) ) j = 1 , j k ( s ) N j = j = 1 , j k ( s ) N b s j j f s k ( s ) j = 1 , j k ( s ) N j 0 b s + 1 , k ( s ) k ( s ) j = 1 , j k ( s ) N b ˜ s + 1 , j j j f s + 1 , k ( s ) j = 1 , j k ( s ) N j b H k ( s ) k ( s ) j = 1 , j k ( s ) N b ˜ H j j j f H , k ( s ) j = 1 , j k ( s ) N j J α
Therefore, by setting:
a i 0 s b s k ( s ) k ( i ) = k ( s ) , 1 i s 1 ( b s , k ( s ) Q b i , k ( s ) Q ) = i 1 = s H a i 0 i 1 b i 1 k ( s ) k ( i ) = k ( s ) , 1 i s 1 ( b i 1 k ( s ) Q b i , k ( s ) Q )
for 1 i 0 M , and by setting b i j = b ˜ i j again, we have:
J = a i 0 i 1 b i 1 k ( i 1 ) k ( i ) = k ( i 1 ) , 1 i i 1 1 ( b i 1 , k ( i 1 ) Q b i , k ( i 1 ) Q ) : 1 i 0 M , 1 i 1 s 1 + a i 0 s b s k ( s ) k ( i ) = k ( s ) , 1 i s 1 ( b s , k ( s ) Q b i , k ( s ) Q ) : 1 i 0 M + a i 0 s + 1 a i 0 H b s + 1 j k ( i ) = j , 1 i s ( b s + 1 j Q b i j Q ) b H j k ( i ) = j , 1 i s ( b H j Q b i j Q ) : j = 1 , , N + a i 0 s + 1 a i 0 H b s + 1 , 1 1 b s + 1 , N N b H 1 1 b H N N : 1 i 0 M , i = 1 N i = n Q + 1 , n 1 , j < Q n + 1
with (i)∼(iv).
Step 2
By Step 1, we need to consider the ideal:
a i 0 , i 1 i = 1 i 1 v i ( c o u n t ( i , i 1 , k ( i 1 ) ) 1 ) Q + 1 : 1 i 0 M , i 1 s + a i 0 s + 1 a i 0 H i = 1 s v i c o u n t ( i , s , j ) Q + 1 b s + 1 j b H j : j = 1 , , N + a i 0 s + 1 a i 0 H i = 1 s v i Q + 1 b s + 1 , 1 1 b s + 1 , N N b H 1 1 b H N N : 1 i 0 M , i = 1 N i = Q + 1 , j < Q + 1
with Jacobian:
i = 1 H v i ( H i + 1 ) N 1 + d i ( s ) d v i k = 1 M i = 1 H d a k i i = 1 H j = 1 N d b i j
where:
d i ( s ) = ( N 1 ) Q s 1 = i s ( c o u n t ( i , s 1 , k ( s 1 ) ) 1 )
We have: λ 0 ( | | A M , H B H , N ( Q ) | | 2 )
min { ( H i + 1 ) N + d i ( s ) + d i ( s ) + d i , i s + 1 , , i H ( s ) 2 ( c o u n t ( i , s , k ( s ) ) 1 ) Q + 2 : 1 i s , i α 0 , 1 s H }
where:
d i ( s ) = M ( i 1 ) { ( c o u n t ( i , s , k ( s ) ) 1 ) Q + 1 } + Q M s 1 = i , c o u n t ( i , s , k ( s ) ) > c o u n t ( i , s 1 , k ( s 1 ) ) s 1 ( c o u n t ( i , s , k ( s ) ) c o u n t ( i , s 1 , k ( s 1 ) ) ) d i , i s + 1 , , i H ( s ) = α = s + 1 H { M i α + j = 1 , j k ( s ) c o u n t ( i , s , j ) = 0 max { Q ( c o u n t ( i , s , k ( s ) ) 1 ) i α , 0 } + j = 1 , j k ( s ) c o u n t ( i , s , j ) 1 , c o u n t ( i , s , k ( s ) ) 2 N max { Q ( c o u n t ( i , s , k ( s ) ) 2 ) i α , 0 } }
Set C ( i , s ) = # { c o u n t ( i , s , j ) = 0 , 1 j N } . Then, we have:
d i ( s ) = min { d i , i s + 1 , , i H , i α 0 } = 0 , if  c o u n t ( i , s , k ( s ) ) = 1 ( H s ) { C ( i , s ) Q + ( N 1 ) Q ( c o u n t ( i , s , k ( s ) ) 2 ) } if  c o u n t ( i , s , k ( s ) ) 2 , N 1 M ( H s ) { C ( i , s ) Q + M Q ( c o u n t ( i , s , k ( s ) ) 2 ) } if  c o u n t ( i , s , k ( s ) ) 2 , C ( i , s ) M < N 1 ( H s ) { M Q ( c o u n t ( i , s , k ( s ) ) 1 ) } if  c o u n t ( i , s , k ( s ) ) 2 , M C ( i , s )
By the above equation, we have bound 1 . By [19], we have bound 2 and bound 3 , thus completing the proof.

References

  1. Hartigan, J.A. A Failure of Likelihood Ratio Asymptotics for Normal Mixtures. In Proceedings of the Berkeley Conference in Honor of J.Neyman and J.Kiefer, California, CA, USA, 1985; Volume 2, pp. 807–810.
  2. Sussmann, H.J. Uniqueness of the weights for minimal feed-forward nets with a given input-output map. Neural Netw. 1992, 5, 589–593. [Google Scholar] [CrossRef]
  3. Hagiwara, K.; Toda, N.; Usui, S. On the problem of applying AIC to determine the structure of a layered feed-forward neural network. In Proceedings of the IJCNN Nagoya Japan, Nagoya Congress Center, Japan, 25–29 October 1993; Volume 3, pp. 2263–2266.
  4. Fukumizu, K. A regularity condition of the information matrix of a multilayer perceptron network. Neural Netw. 1996, 9, 871–879. [Google Scholar] [CrossRef]
  5. Akaike, H. A new look at the statistical model identification. IEEE Trans. Autom. Control 1974, 19, 716–723. [Google Scholar] [CrossRef]
  6. Watanabe, S. Algebraic analysis for nonidentifiable learning machines. Neural Comput. 2001, 13, 899–933. [Google Scholar] [CrossRef] [PubMed]
  7. Watanabe, S. Algebraic geometrical methods for hierarchical learning machines. Neural Netw. 2001, 14, 1049–1060. [Google Scholar] [CrossRef]
  8. Watanabe, S. Algebraic geometry of learning machines with singularities and their prior distributions. J. Jpn. Soc. Artif. Intell. 2001, 16, 308–315. [Google Scholar]
  9. Watanabe, S. Algebraic Geometry and Statistical Learning Theory; Cambridge University Press: New York, NY, USA, 2009; Volume 25. [Google Scholar]
  10. Fulton, W. Introduction to Toric Varieties, Annals of Mathematics Studies; Princeton University Press: Princeton, NJ, USA, 1993. [Google Scholar]
  11. Kollár, J. Singularities of Pairs. In Algebraic Geometry-Santa Cruz 1995, Series Proceedings of Symposia in Pure Mathematics, 9–29 July 1995; American Mathematical Society: Providence, RI, USA, 1997; Volume 62, pp. 221–287. [Google Scholar]
  12. Mustata, M. Singularities of pairs via jet schemes. J. Am. Math. Soc. 2002, 15, 599–615. [Google Scholar] [CrossRef]
  13. Yamazaki, K.; Aoyagi, M.; Watanabe, S. Asymptotic analysis of Bayesian generalization error with Newton diagram. Neural Netw. 2010, 23, 35–43. [Google Scholar] [CrossRef] [PubMed]
  14. Aoyagi, M.; Nagata, K. Learning coefficient of generalization error in Bayesian estimation and Vandermonde matrix type singularity. Neural Comput. 2012, 24, 1569–1610. [Google Scholar] [CrossRef] [PubMed]
  15. Aoyagi, M.; Watanabe, S. Stochastic complexities of reduced rank regression in Bayesian estimation. Neural Netw. 2005, 18, 924–933. [Google Scholar] [CrossRef] [PubMed]
  16. Aoyagi, M.; Watanabe, S. Resolution of singularities and the generalization error with Bayesian estimation for layered neural network. IEICE Trans. J88-D-II 2005, 10, 2112–2124. [Google Scholar]
  17. Aoyagi, M. The zeta function of learning theory and generalization error of three layered neural perceptron. RIMS Kokyuroku Recent Top. Real Complex Singul. 2006, 1501, 153–167. [Google Scholar]
  18. Aoyagi, M. A Bayesian learning coefficient of generalization error and Vandermonde matrix-type singularities. Commun. Stat. Theory Methods 2010, 39, 2667–2687. [Google Scholar] [CrossRef]
  19. Aoyagi, M. Learning coefficient in Bayesian estimation of restricted Boltzmann machine. J. Algebr. Stat. 2013, in press. [Google Scholar] [CrossRef]
  20. Rusakov, D.; Geiger, D. Asymptotic Model Selection for Naive Bayesian Networks. In Proceedings of the Eighteenth Conference on Uncertainty in Artificial Intelligence, Alberta, Canada, 1–4 August 2002; pp. 438–445.
  21. Rusakov, D.; Geiger, D. Asymptotic model selection for naive Bayesian networks. J. Mach. Learn. Res. 2005, 6, 1–35. [Google Scholar]
  22. Zwiernik, P. An asymptotic behavior of the marginal likelihood for general Markov models. J. Mach. Learn. Res. 2011, 12, 3283–3310. [Google Scholar]
  23. Watanabe, S. Equations of states in singular statistical estimation. Neural Netw. 2010, 23, 20–34. [Google Scholar] [CrossRef] [PubMed]
  24. Aoyagi, M. Log canonical threshold of Vandermonde matrix type singularities and generalization error of a three layered neural network. Int. J. Pure Appl. Math. 2009, 52, 177–204. [Google Scholar]
  25. Drton, M. Conference Lecture: Reduced Rank Regression. Workshop on Singular Learning Theory, AIM 2011. Available online: http://math.berkeley.edu/critch/slt2011/ (accessed on 16 December 2011).
  26. Drton, M. Conference Lecture: Bayesian Information Criterion for Singular Models. Algebraic Statistics 2012 in the Alleghenies at The Pennsylvania State University. Available online: http://jasonmorton.com/aspsu2012/ (accessed on 15 June 2012).
  27. Nagata, K.; Watanabe, S. Exchange Monte Carlo Sampling from Bayesian posterior for singular learning machines. IEEE Trans. Neural Netw. 2008, 19, 1253–1266. [Google Scholar] [CrossRef]
  28. Nagata, K.; Watanabe, S. Asymptotic behavior of exchange ratio in exchange Monte Carlo method. Int. J. Neural Netw. 2008, 21, 980–988. [Google Scholar] [CrossRef] [PubMed]
  29. Bernstein, I.N. The analytic continuation of generalized functions with respect to a parameter. Funct. Anal. Appl. 1972, 6, 26–40. [Google Scholar]
  30. Bjőrk, J.E. Rings of Differential Operators; North-Holland: Amsterdam, The Netherlands, 1979. [Google Scholar]
  31. Kashiwara, M. B-functions and holonomic systems. Invent. Math. 1976, 38, 33–53. [Google Scholar] [CrossRef]
  32. Lin, S. Asymptotic approximation of marginal likelihood integrals. 2010; arXiv:1003.5338v2. [Google Scholar]

Share and Cite

MDPI and ACS Style

Aoyagi, M. Consideration on Singularities in Learning Theory and the Learning Coefficient. Entropy 2013, 15, 3714-3733. https://doi.org/10.3390/e15093714

AMA Style

Aoyagi M. Consideration on Singularities in Learning Theory and the Learning Coefficient. Entropy. 2013; 15(9):3714-3733. https://doi.org/10.3390/e15093714

Chicago/Turabian Style

Aoyagi, Miki. 2013. "Consideration on Singularities in Learning Theory and the Learning Coefficient" Entropy 15, no. 9: 3714-3733. https://doi.org/10.3390/e15093714

APA Style

Aoyagi, M. (2013). Consideration on Singularities in Learning Theory and the Learning Coefficient. Entropy, 15(9), 3714-3733. https://doi.org/10.3390/e15093714

Article Metrics

Back to TopTop