Next Article in Journal
On Some Forgotten Formulas of L. de Broglie and the Nature of Thermal Time
Previous Article in Journal
An Optimized Encryption Storage Scheme for Blockchain Data Based on Cold and Hot Blocks and Threshold Secret Sharing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Efficient First-Order Algorithms for Large-Scale, Non-Smooth Maximum Entropy Models with Application to Wildfire Science

by
Gabriel Provencher Langlois
1,*,
Jatan Buch
2 and
Jérôme Darbon
3
1
Courant Institute of Mathematical Sciences, New York University, New York, NY 10012, USA
2
Department of Earth and Environmental Engineering, Columbia University, New York, NY 10027, USA
3
Division of Applied Mathematics, Brown University, Providence, RI 02912, USA
*
Author to whom correspondence should be addressed.
Entropy 2024, 26(8), 691; https://doi.org/10.3390/e26080691
Submission received: 8 May 2024 / Revised: 17 July 2024 / Accepted: 6 August 2024 / Published: 15 August 2024

Abstract

:
Maximum entropy (MaxEnt) models are a class of statistical models that use the maximum entropy principle to estimate probability distributions from data. Due to the size of modern data sets, MaxEnt models need efficient optimization algorithms to scale well for big data applications. State-of-the-art algorithms for MaxEnt models, however, were not originally designed to handle big data sets; these algorithms either rely on technical devices that may yield unreliable numerical results, scale poorly, or require smoothness assumptions that many practical MaxEnt models lack. In this paper, we present novel optimization algorithms that overcome the shortcomings of state-of-the-art algorithms for training large-scale, non-smooth MaxEnt models. Our proposed first-order algorithms leverage the Kullback–Leibler divergence to train large-scale and non-smooth MaxEnt models efficiently. For MaxEnt models with discrete probability distribution of n elements built from samples, each containing m features, the stepsize parameter estimation and iterations in our algorithms scale on the order of O ( m n ) operations and can be trivially parallelized. Moreover, the strong 1 convexity of the Kullback–Leibler divergence allows for larger stepsize parameters, thereby speeding up the convergence rate of our algorithms. To illustrate the efficiency of our novel algorithms, we consider the problem of estimating probabilities of fire occurrences as a function of ecological features in the Western US MTBS-Interagency wildfire data set. Our numerical results show that our algorithms outperform the state of the art by one order of magnitude and yield results that agree with physical models of wildfire occurrence and previous statistical analyses of wildfire drivers.

1. Introduction

Maximum entropy (MaxEnt) models are a class of density estimation methods that use the maximum entropy principle [1] to reproduce key statistics of data sets. Historically used in physics [1,2], engineering [3,4,5], and statistics [6] applications, MaxEnt models are now frequently used for large-scale machine learning problems in natural language processing [7,8,9,10,11,12], social science [13,14,15], neuroscience [16,17,18], ecological modeling [19,20,21,22,23,24,25,26,27,28], and wildfire science [29,30,31]. MaxEnt models in machine learning must often estimate probability distributions from big data sets comprising hundreds of thousands to billions of samples or features or both [32]. Due to this enormous number, large-scale MaxEnt models need efficient and robust algorithms to perform well.
State-of-the-art algorithms for MaxEnt models, however, were not originally designed to handle big data sets; these algorithms either rely on technical devices that may yield unreliable numerical results [33], scale poorly in size [9,34], or require smoothness assumptions that many MaxEnt models lack in practice [10]. These limitations make it essentially impossible to scale MaxEnt models to big data applications without adequate and costly computational resources [35,36]. This constraint on computational resources, in particular, has been recently identified as a crucial challenge to overcome for using large-scale MaxEnt models in climate change and wildfire studies within a reasonable amount of run time [27,28,30,37].

1.1. Contributions of This Paper

In this paper, we present novel optimization algorithms that overcome the shortcomings of state-of-the-art algorithms used for training large-scale, non-smooth MaxEnt models. Our proposed algorithms are first-order accelerated nonlinear primal–dual hybrid gradient (NPDHG) algorithms, whose theory was provided by two authors of this paper in [38], based on the Kullback–Leibler divergence. Using the Kullback–Leibler divergence over the classical proximal operator makes it possible to train large-scale and non-smooth MaxEnt models much more efficiently than the state of the art. In particular, we show that, for a MaxEnt model with a discrete probability distribution of n elements built from samples each containing m features, the stepsize parameter estimation and iterations in our algorithms all scale on the order of O ( m n ) operations. This significantly improves on the known complexity bound of O ( min ( m 2 n , m n 2 ) ) operations for computing the optimal stepsize parameters of classical first-order optimization methods, such as the linear PDHG or forward–backward splitting methods. We also show, as a consequence, that, for a given tolerance level ϵ > 0 , our algorithms provably compute solutions using on the order of O ( m n / ϵ ) or O ( m n / log ( 1 / ϵ ) ) operations, the order depending on the smoothness of the MaxEnt model and which are optimal with respect to the Nesterov class of optimal first-order methods [39]. Moreover, the computational bottleneck consists of matrix–vector multiplications, which can be trivially parallelized, and so our proposed algorithms exhibit scalable parallelism. Finally, we show that the strong convexity of the Kullback–Leibler divergence with respect to the 1 norm naturally allows for significantly larger stepsize parameters, thereby substantially speeding up the convergence rate of our algorithms.
To illustrate the efficiency of our novel algorithms on large-scale problems, we present an application to wildfire science. Specifically, we consider the problem of estimating the probabilities of fire occurrences across the western US as a function of ecological features. To do so, we fit elastic net, non-overlapping group lasso, and MaxEnt models to a large number of hyperparameters using our proposed algorithms and the state-of-the-art forward–backward splitting and coordinate descent STRUCTMAXENT2 algorithms [40,41]. Our numerical results show that our algorithms outperform the two latter algorithms by at least one order of magnitude, and yield results that are in good agreement with physical models of wildfire occurrence [42,43] as well as previous statistical analyses of wildfire drivers [29,30,31,44].

1.2. Organization of This Paper

This paper is organized as follows. In Section 2, we present our setup, we describe how MaxEnt models work, and we review three popular non-smooth MaxEnt models that are challenging to train with big data sets: the elastic net, group lasso, and regularized MaxEnt models. In Section 3, we explain why large-scale, non-smooth MaxEnt models are particularly challenging to train from big data sets, and we describe the state-of-the-art algorithms for training these MaxEnt models and their limitations. In Section 4, we describe our approach for estimating MaxEnt models with NPDHG optimization methods, derive explicit algorithms for regularized MaxEnt models, and explain why our algorithms overcome the limitations of state-of-the-art algorithms. In Section 5, we present an application of our algorithms to wildfire science, where we train a large number of MaxEnt models on the Western US MTBS-Interagency (WUMI) wildfire data set [45] to estimate the probabilities of fire occurrences as a function of ecological features. Finally, we review our results and outline directions for future work in Section 6.

2. Preliminaries

2.1. Setup

Suppose we receive l independent, identically distributed samples { v 1 , , v l } I from an unknown distribution D . We assume throughout, without of loss of generality, that the input space I { 1 , , n } . In addition, suppose we have a prior probability distribution p prior on I that encapsulates some prior knowledge about the samples or unknown distribution. Finally, suppose we have access to a set of features from the samples via a bounded feature map Φ : I R m . Then, how do we estimate the unknown distribution D from the prior distribution p prior , the samples { v 1 , , v l } , and the feature map Φ ?
The maximum entropy principle offers a systematic way to answer this question. It states that the distribution that best estimates the unknown distribution D is the one that remains as close as possible to the prior probability p prior while matching the features { Φ ( v 1 ) , , Φ ( v l ) } exactly or as closely as possible, in some suitable sense. Specifically, let Δ n denote the n-dimensional probability simplex and suppose p prior int Δ n (i.e., p prior ( j ) > 0 for every j { 1 , , n } ). We measure the closeness of a probability distribution p Δ n to the prior probability p prior using the Kullback–Leibler divergence D KL : Δ n × int Δ n R :
D KL ( p p prior ) = j = 1 n p ( j ) log p ( j ) / p prior ( j ) .
Next, let D ^ denote the empirical distribution induced by samples { v 1 , , v l } :
D ^ ( j ) = 1 l { 1 i l v i = j } .
In addition, let E D ^ [ Φ ] denote the empirical average of the features induced by the samples and let E p [ Φ ] denote the model average induced by a probability distribution p Δ n :
E D ^ [ Φ ] = j = 1 n D ^ ( j ) Φ ( j ) and E p [ Φ ] = j = 1 n p ( j ) Φ ( j ) .
We measure how the averages E D ^ [ Φ ] and E p [ Φ ] match using a function H * : R m R { + } called the potential function. MaxEnt models combine the two measures described above and seek to minimize the sum
inf p Δ n f ( p ; t ) = inf p Δ n D KL ( p p prior ) + t H * E D ^ [ Φ ] E p [ Φ ] t
where t > 0 is a hyperparameter selected using data-driven methods, e.g., cross-validation.
Standing assumptions: The potential function is proper, lower semicontinuous, convex, and bounded from below by zero with H * ( 0 ) = 0. Under the standing assumptions above on the potential function, which are fairly general, the probability distribution D ^ is a feasible point of (3). Moreover, the MaxEnt estimation problem (3) admits a unique global solution (see ([46], Page 35, Proposition 1.2); uniqueness follows from the strong convexity of the Kullback–Leibler divergence with respect to the 1 norm [47,48,49,50,51]).

2.2. Dual Formulation and Optimality Conditions

The generalized MaxEnt problem (3) admits a dual problem corresponding to regularized maximum a posteriori estimation [52]. To derive the dual problem, we first write the second term on the right hand side of (3) in terms of its convex conjugate,
t H * E D ^ [ Φ ] E p [ Φ ] t = sup w R m w , E D ^ [ Φ ] E p [ Φ ] t H ( w ) ,
where we abuse notation and write the convex conjugate of H * as H. This lets us express problem (3) in saddle-point form:
inf p Δ n sup w R m w , E D ^ [ Φ ] E p [ Φ ] t H ( w ) + D KL ( p p prior ) .
Due to our assumptions on the potential function H * , we can swap the infimum and supremum ([46], Statement (4.1) on page 61). Using the convex conjugate formula
inf p Δ n D KL ( p p prior ) w , E p [ Φ ] = log j = 1 n p prior ( j ) e w , Φ ( j ) ,
we obtain the dual problem of (3)
sup w R m w , E D ^ [ Φ ] t H ( w ) log j = 1 n p prior ( j ) e w , Φ ( j ) .
The dual problem (5) is a regularized maximum likelihood estimation problem over the family of Gibbs distributions [41,52]. It has at least one global solution ([46], Proposition (1.2) on page 35). Moreover, if ( p s , w s ) denotes a pair of solutions to the MaxEnt estimation problem (3) and its dual (5), then this pair satisfies the optimality conditions
E D ^ [ Φ ] E p s [ Φ ] t H ( w s ) and p s ( j ) = p prior ( j ) e w s , Φ ( j ) j = 1 n p prior ( j ) e w s , Φ ( j ) ,
for every j { 1 , , n } where H ( w s ) denotes the subdifferential of H at w s .

2.3. Examples of Non-Smooth MaxEnt Models

MaxEnt models differ in the choice of the prior distribution p prior , hyperparameter t, and the potential function H * . In practice, the prior distribution is often chosen to be uniform and the hyperparameter is chosen via data-driven methods. The choice of potential function depends on the application. The classical MaxEnt model uses the indicator function
u H * ( u ) = { 0 , if u = 0 , + otherwise .
This forces the model average E p [ Φ ] to be equal to the empirical average E D ^ [ Φ ] . However, this is often too restrictive because the averages are expected to be close, and not equal, with high probability. Forcing the model and empirical averages to be equal can lead to severe over-fitting issues with big data sets, and more flexibility is therefore desired.
We review below three different non-smooth MaxEnt models that offer this flexibility: the elastic net, group lasso, and -regularized MaxEnt models. We chose these models because they are used extensively in statistics and machine learning but are also challenging to train on big data sets due to their non-smoothness [20,26,27,53]. These models will also be used for our numerical experiments and results in Section 5.

2.3.1. Elastic Net Regularized MaxEnt Models

These models use for potential function the relaxation of the convex conjugate of the 1 -norm:
u H * ( u ) 1 α 2 · 2 2 + α · 1 * ( u ) = 1 2 ( 1 α ) i = 1 m max ( 0 , | u i | α ) 2
where α ( 0 , 1 ) . This yields the elastic net regularized maximum entropy estimation problem
min p Δ n D KL p ( p p prior ) + 1 2 t ( 1 α ) i = 1 m max 0 , [ E D ^ [ Φ ] ] i [ E p [ Φ ] ] i t α 2 .
This type of regularization is frequently used for feature selection in machine learning [12,20,54,55]. The corresponding dual problem is
sup w R m w , E D ^ [ Φ ] t ( 1 α ) 2 w 2 2 + α w 1 log j = 1 n p prior ( j ) e w , Φ ( j ) .
Thanks to the elastic net penalty ( 1 α ) 2 w 2 2 + α w 1 , problem (8) is strongly concave and has a unique solution.
An important aspect of the elastic net penalty is that it promotes sparsity; that is, it leads to a solution with many entries equal to zero [56,57,58], the number depending on the hyperparameter t and parameter α . Sparsity is useful when seeking accurate solutions based only on a few features or when performing feature selection.

2.3.2. Group Lasso Regularized MaxEnt Models

These models are specified starting from the dual problem. Consider a partition G of the components of a vector w R m into g different and possibly overlapping groups [ w 1 , , w g ] with w g R m g and g = 1 G w g = w . Let w w g 2 , g denote the 2 norm over the respective components of the gth group. Then, the dual version of the group lasso regularized maximum entropy estimation problem is
sup w R m w , E D ^ [ Φ ] t g = 1 G m g w g 2 , g log j = 1 n p prior ( j ) e w , Φ ( j ) .
Thanks to the group lasso penalty g = 1 G m g w g 2 , g , problem (10) is strictly concave and so has at least one global solution. The corresponding primal problem follows from the convex conjugate formula
u H * ( u ) g = 1 G m g · 2 , g * ( u ) = { 0 , if u g 2 , g m g for g = 1 , , G , + , otherwise ,
and it is therefore given by
min p Δ n D KL ( p p prior ) s . t . E D ^ [ Φ g ] E p [ Φ g ] 2 , g t m g for g = 1 , , G .
Similarly to the elastic net penalty, the group lasso penalty promotes sparsity in the global solutions of (9), but it differs in that it sets different blocks of components [ w 1 , , w g ] to zero at the same time.

2.3.3. -Regularized MaxEnt Models

These models use for potential function the convex conjugate of the norm,
u H * ( u ) u * = u ^ R m u u ^ 1 1 ,
which is the characteristic set of the 1 ball with unit radius. This yields the -regularized maximum entropy estimation problem
min p Δ n D KL ( p p prior ) + t E D ^ [ Φ ] E p [ Φ ] t * .
This type of regularization is used for certain machine learning problems, including matrix factor decomposition, and in robust statistics applications (see, e.g., [53,59]). The corresponding dual problem is
sup w R m w , E D ^ [ Φ ] t w log j = 1 n p prior ( j ) e w , Φ ( j )
Thanks to the norm, problem (12) is strictly concave and has at least one global solution.

3. Related Work

3.1. Large-Scale Sparse MaxEnt Models: Computational Challenges

Estimating a probability distribution from the MaxEnt model (3) can be computationally prohibitive for big data sets. To illustrate this point, suppose the MaxEnt model (3) with hyperparameter t 0 has a global solution p s ( t ) . Let p ϵ ( t ) Δ n with ϵ > 0 denote an ϵ -approximate solution to the global solution p s ( t ) , that is, the objective function f in (3) satisfies
f ( p ϵ ( t ) ; t ) f ( p s ( t ) ; t ) < ϵ .
If the potential function H * is smooth (equivalently, its conjugate H is strongly convex), the best achievable rate of convergence for computing p ϵ ( t ) with t > 0 in the Nesterov class of optimal first-order methods is linear O ( log ( 1 / ϵ ) ) in the number of operations [39]. If H is not smooth, then the optimal convergence rate is sublinear O ( 1 / ϵ ) in the number of operations.
These rates, while optimal, require carefully fine-tuned stepsize parameters. In classical first-order optimization algorithms, these stepsize parameters are fine-tuned using a precise estimate of the largest singular value of the linear operator A : Δ n R m defined by
A p = j = 1 n p ( j ) Φ ( j ) = E p [ Φ ] .
Unfortunately, computing the largest singular value of the linear operator A in (13) accurately is often computationally expensive for large matrices due to its prohibitive cubic cost of O ( min ( m 2 n , m n 2 ) ) operations [60]. In this situation, line search methods and other heuristics can sometime be employed to bypass this issue, but they typically slow down the convergence speed. Another approach is to compute a crude estimate of the largest singular value of the matrix A , but doing so significantly reduces convergence speed as well. In fact, even if the largest singular value can be computed quickly, the resulting stepsize parameters may be much smaller than what is permissible to maintain convergence, that is, the largest singular value itself may be a poor estimate for determining how large the stepsize parameters are allowed to be to maintain convergence. This point will discussed in more detail in Section 4.1 when we describe our proposed NPDHG algorithms.
This issue makes solving the MaxEnt model (3) difficult and laborious. Even worse, in some applications, the appropriate value of the hyperparameter t in (3) is difficult to guess and must be selected by repeatedly solving (3) from a large pool of values of hyperparameters, a process that can become particularly time-consuming and resource intensive for big data sets. This issue has driven much research in developing robust and efficient algorithms to minimize computational costs and maximize model performance.

3.2. State-of-the-Art Methods for Large-Scale, Non-Smooth MaxEnt Models

State-of-the-art methods for computing solutions to large-scale, non-smooth MaxEnt models are based on coordinate descent algorithms [61,62,63,64] and first-order optimization algorithms such as forward–backward splitting [40,65,66]. We will discuss these methods below, but note that other types of methods have been developed for MaxEnt models; see, e.g., [10,41,67,68] for surveys and comparisons of different algorithms. In particular, we will not consider second-order-based methods suitable for smooth MaxEnt models such as limited-memory BFGS algorithms [10,69] since this work focuses on non-smooth MaxEnt models.
Coordinate descent methods. The state of the art is a coordinate descent algorithm based on a technical device called infinitely weighted logistic regression (IWLR) [26,33]. The IWLR method approximates a MaxEnt model as a logistic regression model and then fits the approximate logistic regression model using an existing, efficient optimization algorithm for logistic regression. This indirect approach was proposed because efficient and scalable coordinate descent algorithms were already available for fitting logistic regression models, which in [33] was an earlier version of the GLMNET software package [63]. The IWLR method remains the state of the art for this reason. It is implemented, for example, in the MaxEnt package available in the R programming language [26,63].
The IWLR method, however, is an approximate technical device that is not guaranteed to work, a fact acknowledged by the authors who proposed the method [33], and therefore may not produce reliable numerical results. Coordinate descent algorithms themselves are generally non-parallelizable and often lack robustness and good convergence properties. For example, the aforementioned GLMNET software package approximates the logarithm term in logistic regression models with a quadratic term to fit the models efficiently. Without costly stepsize optimization, which GLMNET avoids to improve performance, the GLMNET implementation may not converge [62,70]. A case in point is that [71] provides two numerical experiments in which GLMNET does not converge.
Other coordinate descent algorithms have been developed to compute solutions to non-smooth MaxEnt models directly (see, e.g., [64,72]) but they are also generally non-parallelizable and often lack robustness and good convergence properties. Finally, another issue is that many coordinate descent algorithms, including GLMNET, were designed for sparse MaxEnt models (e.g., the elastic net and group lasso regularized MaxEnt models described in Section 2.3) and those algorithms depend on the sparsity of the model to converge quickly [54]. It would be desirable to have fast optimization methods for when the feature mapping (13) is dense, as this often occurs in practice.
First-order methods. First-order optimization algorithms such as the forward–backward splitting algorithm are popular because they are robust and can provably compute ϵ -approximate solutions of (3) with an optimal rate of convergence. However, as discussed in detail at the end of the previous subsection, achieving this rate of convergence requires fine-tuning the stepsize parameters using an accurate estimate of the largest singular value of the feature mapping (13), and this estimate is typically computationally expensive for large matrices. This problem makes classical first-order optimization algorithms inefficient and impractical in estimating probability densities from large-scale MaxEnt models.
Summary. The current state-of-the-art algorithms for estimating probability densities from MaxEnt models either produce unreliable numerical results, lack scalable parallelism, or scale poorly in size. These shortcomings in terms of robustness and efficiency make it challenging to use non-smooth, large-scale MaxEnt models in applications without access to adequate and costly computational resources. The next section presents novel, efficient, and robust accelerated NPDHG optimization methods that address these shortcomings.

4. Main Results

We describe here our approach for computing solutions to the generalized MaxEnt problem (3) using accelerated NPDHG optimization methods [38]. In addition to the standing assumptions on the potential function H * , we assume we can compute efficiently the proximal operator of the convex conjugate of the potential function
arg min w R m λ H ( w ) + 1 2 w w ^ 2 2 ,
for every λ > 0 and w ^ R m . These assumptions are satisfied for most potential functions used in practice, including for the potential functions described in Section 2.3.

4.1. Methodology

We start with the saddle-point formulation (4) of the generalized MaxEnt estimation problem (3). Based on the work two of the authors provided in [38], we propose to split the infimum and supremum in the saddle-point problem (4) using an iterative NPDHG scheme that alternates between a nonlinear proximal ascent step based on the Kullback–Leibler divergence and a linear proximal descent step.
More precisely, let τ 0 > 0 and σ 0 > 0 be two stepsize parameters satisfying the inequality
τ 0 σ 0 max j { 1 , , n } Φ ( j ) 2 2 1 ,
let θ 0 = 0 , let w 1 = w 0 int dom H , let z 0 R m , and define the initial probability distribution p 0 Δ n through the constant z 0 via
p 0 ( j ) = p prior ( j ) e z 0 , Φ ( j ) j = 1 n p prior ( j ) e z 0 , Φ ( j )
for each j { 1 , , n } . Our proposed NPDHG iterative scheme computes the iterates
p k + 1 = arg min p Δ n τ k D KL ( p p prior ) τ k w k + θ k ( w k w k 1 ) , E p [ Φ ] + D KL ( p p k ) w ^ k + 1 = w k + σ k ( E D ^ [ Φ ] E p k + 1 [ Φ ] ) , w k + 1 = arg min w R m t σ k H ( w ) + 1 2 w w ^ k + 1 2 2 , θ k + 1 = 1 / 1 + τ k , τ k + 1 = θ k + 1 τ k and σ k + 1 = σ k / θ k + 1 .
According to ([38], Proposition 4.1), the sequence of iterates p k converges strongly to the unique solution of the generalized MaxEnt problem (3). Moreover, for any t 0 and a given tolerance level ϵ > 0 , the scheme (16) provably computes an ϵ -approximate solution p ϵ ( t ) of the generalized MaxEnt model (3) in O ( 1 / ϵ ) time. This rate of convergence is, without further smoothness assumptions on the potential function H * , the best achievable rate of convergence with respect to the Nesterov class of optimal first-order methods [39].
The key element in this scheme is the choice of the Kullback–Leibler divergence as a nonlinear proximal step in the first line of (16). We use it for two reasons: First, because the Kullback–Leibler divergence already appears in the saddle-point problem (4). This allows us to compute p k + 1 explicitly. Indeed, thanks to the choice of initial probability distribution (15), we have
p k + 1 ( j ) = p prior ( j ) e z k + 1 , Φ ( j ) j = 1 n p prior ( j ) e z k + 1 , Φ ( j ) with z k + 1 = 1 1 + τ k z k + τ k ( w k + θ k ( w k w k 1 ) )
for each j { 1 , , n } . See Appendix A for the derivation of (17).
Second, because the Kullback–Leibler divergence is 1-strongly convex with respect to the 1 norm, that is,
D KL ( p p k ) 1 2 p p k 1 2 .
This fact follows from a fundamental result in information theory known as Pinsker’s inequality [47,48,49,50,51]. In particular, this means that scheme (16) alternates between solving a strongly convex problem over the space ( R n , · 1 ) and a concave problem over the space ( R m , · 2 ) . The choice of these spaces is significant, for the induced operator norm · op of the linear operator A defined in (13) becomes
A op = sup p 1 = 1 A p 2 = max j { 1 , , n } Φ ( j ) 2 .
This operator norm offers two crucial advantages: First, it can be computed in optimal Θ ( m n ) time, or better if the features { Φ ( j ) } j = 1 n have structure, e.g., if they are sparse. This means that the stepsize parameters τ 0 and σ 0 of the NPDHG scheme can be computed in (14), with equality, in optimal Θ ( m n ) time. This is important because, in classical first-order optimization methods, we typically require a precise estimate of the largest singular value of the feature mapping (13), namely the number
A 2 = sup x 2 = 1 A p 2 ,
to fine-tune the stepsize parameters to gain an optimal convergence rate. However, as discussed in detail in Section 3.1, this estimate is computationally expensive for large matrices due to its prohibitive cubic cost of O ( min ( m 2 n , m n 2 ) ) operations. In contrast, our NPDHG scheme (16) does not suffer from this computational bottleneck and therefore scales well. Second, the operator norm A op can be significantly smaller than the estimate A 2 , hence allowing for bigger stepsize parameters to further speed up convergence. (An easy calculation yields max p 2 = 1 A p 2 max j { 1 , , n } Φ ( j ) 2 .)
Summary. To solve the generalized MaxEnt estimation problem (3) and its dual problem (5), let τ 0 > 0 and σ 0 > 0 be two stepsize parameters satisfying inequality (14), let θ 0 = 0 , let w 1 = w 0 int dom H , let z 0 R m , and define an initial probability distribution p 0 through z 0 via (15). Then, compute the iterates
z k + 1 = z k + τ k ( w k + θ k ( w k w k 1 ) ) / ( 1 + τ k ) , p k + 1 ( j ) = p prior ( j ) e z k + 1 , Φ ( j ) j = 1 n p prior ( j ) e z k + 1 , Φ ( j ) for j { 1 , , n } , w ^ k + 1 = w k + σ k ( E D ^ [ Φ ] E p k + 1 [ Φ ] ) , w k + 1 = arg min w R m t σ k H ( w ) + 1 2 w w ^ k + 1 2 2 , θ k + 1 = 1 / 1 + τ k , τ k + 1 = θ k + 1 τ k and σ k + 1 = σ k / θ k + 1 ,
until convergence is achieved. As all parameters and updates can be computed in O ( m n ) time, for any t 0 and a given tolerance level ϵ > 0 , the overall complexity for computing an ϵ -approximate solution p ϵ ( t ) is O ( m n / ϵ ) .

4.2. Algorithm for Smooth Potential Functions

The iterative scheme (18) does not require the potential function H * to be smooth. If, however, the potential function H * is γ H * -smooth (equivalently, H is 1 γ H * -strongly convex) for some γ H * > 0 , then we can modify the NPDHG iterative scheme (18) to achieve a linear rate of convergence. More precisely, let t > 0 and introduce the stepsize parameters
θ = 1 t 2 γ H * A op 2 1 + 4 γ H * A op 2 t 1 , τ = 1 θ θ and σ = γ H * τ t .
Let w 1 = w 0 int dom H , let z 0 R m , and define p 0 through z 0 via (15). Then, the explicit NPDHG iterative scheme is
z k + 1 = z k + τ ( w k + θ ( w k w k 1 ) ) / ( 1 + τ ) , p k + 1 ( j ) = p prior ( j ) e z k + 1 , Φ ( j ) j = 1 n p prior ( j ) e z k + 1 , Φ ( j ) for j { 1 , , n } , w ^ k + 1 = w k + σ ( E D ^ [ Φ ] E p k + 1 [ Φ ] ) , w k + 1 = arg min w R m t σ H ( w ) + 1 2 w w ^ k + 1 2 2 .
According to ([38], Proposition 4.3), the sequences of iterates p k and w k converge strongly to the unique solution of the generalized MaxEnt estimation problem (3) and its dual problem (5). Moreover, for any t > 0 and a given tolerance level ϵ > 0 , this scheme provably computes an ϵ -approximate solution p ϵ ( t ) of the generalized MaxEnt estimation problem (3) in O ( log ( 1 / ϵ ) ) . This rate of convergence is the best achievable rate of convergence with respect to the Nesterov class of optimal first-order methods [39]. As all parameters and updates can be computed in O ( m n ) time, the overall complexity for computing an ϵ -approximate solution p ϵ ( t ) is O ( m n log ( 1 / ϵ ) ) .

5. Application to Wildfire Science

To illustrate the efficiency of our novel algorithms on large-scale problems, we present here an application to wildfire science. The problem at hand is to combine fire occurrence data with ecological data in a fixed geographical region to estimate the probability of fire occurrences as a function of ecological features. MaxEnt models achieve this goal by translating fire occurrence and ecological data into probabilities of fire occurrences and ecological features. This approach closely mirrors how MaxEnt models are used for modeling species’ geographic distributions [19,20,22,24,25,26,27,28]. Another related goal is to identify what ecological features correlate most with fire occurrences. This can be achieved using a sparse MaxEnt model, e.g., an elastic net or group lasso regularized MaxEnt model, to identify ecological features correlating significantly with fire occurrences.
For this application, we use the Western US MTBS-Interagency (WUMI) wildfire data set [45], which we describe in Section 5.1 below. We formulate the problem of combining the fire occurrence and ecological data from the WUMI wildfire data set into a MaxEnt estimation problem in Section 5.2. Using this data, we then fit the elastic net, (non-overlapping) group lasso, and MaxEnt models for a large number of hyperparameters weighting the regularization. We detail this fitting procedure and the explicit NPDHG algorithms for these MaxEnt models in Section 5.3. Following this, we compare in Section 5.4 the running times required to fit the aforementioned MaxEnt models using our NPDHG algorithms with the forward–backward splitting algorithm [40,65] and the STRUCTMAXENT2 coordinate descent algorithm of [41]. Finally, in Section 5.5, we interpret the results obtained from fitting the aforementioned MaxEnt models to the WUMI wildfire data set.

5.1. WUMI Wildfire Data Set

The Western US MTBS-Interagency wildfire data set [45] consists of all fires (⩾1 km2) from multiple state and federal agencies, supplemented by satellite observations of large fires (⩾4 km2) from the Monitoring Trends in Burn Severity (MTBS) program, in the continental United States west of 103° W longitude. For this application, we extracted all wildfires from the WUMI data set that occurred between 1984 and 2020 inclusive (accessed 18 May 2023). The locations of all fires used are shown in Figure 1.
Next, following the procedure outlined in [44], we overlay the fire locations on a 12 km × 12 km grid to construct a data frame of prevailing climate, vegetation, topographic, and human-related features for each fire. Altogether we include a total of 35 potential fire-related features from various sources; we provide a summary of all features used in our analysis along with their sources in Appendix B. For each grid cell, we also provide an index identifying its Environmental Protection Agency (EPA) Level III ecoregion. Defined on the basis of roughly homogeneous climate and vegetation patterns, ecoregions are commonly used in the wildfire science literature to identify climate–fire relationships at a coarse spatial scale [29,73]. The full WUMI wildfire data set with associated fire-related features for all months from 1984 to 2020 is publicly available as part of the Stochastic Machine Learning for wildfire activity in the western US (SMLFire1.0) package [44].

5.2. Data Preprocessing

We construct a weakly informative prior [74,75] for incorporating existing knowledge of wildfire conducive environmental conditions in our MaxEnt model. First, we prepare a training data set of 10,000 fire occurrences and absences chosen randomly across all months between 1984 and 2020 correlated with the values of 35 fire-related features described in the previous section. The features are averaged over each calendar month (i.e., January, February, …) between 1984 and 2020. We then apply min–max scaling to each feature, ensuring that all features lie in the same range. Second, we construct two Random Forest (RF) models, one each for fires in dry (May–September) and wet (October–April) seasons respectively. We withhold 20% of the training data to tune the model hyperparameters, such that the optimal hyperparameters ensure a trade-off between model precision and recall. Next, we predict the fire probability for all grid cells with fire-related features using either the trained wet or dry season RF model depending on the calendar month. Since the MaxEnt algorithm assumes a presence-only framework, that is, the absence of fires in a grid cell does not imply a non-zero fire probability, we impute grid cells without a fire with probability p nfire = min ( p fire ) / 10 . Last, we normalize the prior probability distribution to ensure that the fire probability across all grid cells sums up to one. For convenience, we represent the prior distribution as the normalized mean of 12 monthly fire probabilities predicted by the RF models in Figure 1.
To construct the empirical distribution D ^ , we divide the study region into its 18 different EPA level III ecoregions and weigh the relative frequencies of fire among the ecoregions using the following strategy. For each ecoregion r { 1 , , 18 } , let n r , total denote the total number of grid cells in ecoregion r and let n r , fire denote the total number grid cell in ecoregion r where at least one fire was recorded. In addition, let Z = r = 1 18 n r , fire n r , total denote the sum of the proportions of grid cells where at least one fire was recorded among all ecoregions. Then, we compute the empirical probability at cell j { 1 , , n } , D ^ ( j ) , as follows:
D ^ ( j ) = 1 Z 0 if no fire was recorded in grid cell j , 1 / n r , total if at least one fire was recorded in grid j and the grid cell j belongs to ecoregion r
This construction gives the empirical distribution more weight to ecoregions where fires are more frequent and widespread.

5.3. Fitting Procedure and Algorithmic Setup

For the analysis, we fit six different MaxEnt models to the wildfire data: four elastic net MaxEnt models with parameters α = { 0.95 , 0.4 , 0.15 , 0.05 } , a non-overlapping group lasso MaxEnt model, and an -regularized MaxEnt model. For each model, we fit a regularization path of 141 hyperparameters as follows:
t ( l ) = ( 1 l / 100 ) t ( 0 ) from l = 0 to 50 ( 0.5 ( l 50 ) / 200 ) t ( 0 ) from l = 51 to 140 ,
where t ( 0 ) depends on the MaxEnt model and corresponds to the smallest hyperparameter for which the primal and dual solutions are equal to the prior distribution and zero. More precisely, for each model we set ( p ( 0 ) , w ( 0 ) ) = ( p prior , 0 ) and compute the sequence of solutions { ( p ( l ) , w ( l ) ) } l = 1 140 of the corresponding MaxEnt primal and dual problems (7) and (8) with hyperparameter t = t ( l ) .
We chose these MaxEnt models to study the impact of the regularization on the primal and dual solutions as a function of the hyperparameters. In particular, for the elastic net and non-overlapping group lasso MaxEnt models, we are interested in identifying the set of features that are selected or discarded as a function of the sequence of hyperparameters { t ( l ) } l = 0 140 . In the parts below, we describe the value of t ( 0 ) and the NPDHG algorithm used for each model.

5.3.1. Elastic Net MaxEnt Models

For these MaxEnt models, the smallest hyperparameter for which the solutions to the primal and dual problems (7) and (8) are equal to the prior distribution and zero is
t ( 0 ) = E D ^ [ Φ ] E p prior [ Φ ] / α .
This follows from the optimality condition
E D ^ [ Φ ] E p s [ Φ ] t ( 1 α ) w s t α · 1 w s
and computing the smallest parameter t for which (22) is satisfied at ( p s , w s ) = ( p prior , 0 ) .
For the elastic net MaxEnt model with α = 0.95 , we use the NPDHG algorithm (18) (with sublinear convergence rate), while, for the models with α = { 0.4 , 0.15 , 0.05 } , we use the NPDHG algorithm (20) (with linear convergence rate). Starting from l = 1 , we compute the pair of solutions ( p ( l ) , w ( l ) ) to the primal and dual problems (7) and (8) at hyperparameter t = t ( l ) using the previously computed pair of solutions ( p ( l 1 ) , w ( l 1 ) ) by setting the initial vectors z 0 = w 0 = w ( l 1 ) in both NPDHG algorithms. For the stepsize parameters, we set θ 0 = 0 , τ 0 = 2 , and σ 0 = 1 / 2 A op 2 in (18), and we set θ , τ , and σ according to the formulas in (19) with γ H * = 1 α in (20). We compute the update w k + 1 in both NPDHG algorithms using the classical soft thresholding operator [66,76,77]. Specifically, for any λ > 0 , w ^ R m , and i { 1 , , m } we have
arg min w R m λ w 1 + 1 2 w w ^ 2 2 i shrink 1 ( w ^ , λ ) i = { [ w ^ ] i λ if [ w ^ ] i > λ , 0 if | [ w ^ ] i | λ , [ w ^ ] i + λ if [ w ^ ] i < λ ,
and so, for every λ > 0 , α [ 0 , 1 ] , w ^ R m and i { 1 , , m } we have
arg min w R m λ α w 1 + ( 1 α ) 2 w 2 2 + 1 2 w w ^ 2 2 i = [ shrink 1 ( w ^ , λ α ) ] i 1 + λ ( 1 α ) .
Finally, we let the NPDHG algorithms run for at least 40 iterations before checking for convergence. We stop the NPDHG algorithms when the optimality condition (22) is satisfied within some tolerance 10 5 :
E D ^ [ Φ ] E p k [ Φ ] t ( 1 α ) w k t α ( 1 + 10 5 ) .

5.3.2. Non-Overlapping Group Lasso MaxEnt Model

For this MaxEnt model, we divide the features into five disjoint groups of features, as described in Appendix B. Then, the smallest hyperparameter for which the solutions to the primal and dual problems (7) and (8) are equal to the prior distribution and zero is
t ( 0 ) = max g { 1 , , 5 } E D ^ [ Φ g ] E p [ Φ g ] 2 , g / m g ,
where m g is the number of features in the gth group. This follows from the optimality condition
E D ^ [ Φ ] E p s [ Φ ] g = 1 5 t m g · 2 , g w g s
and computing the smallest parameter t for which (23) is satisfied at ( p s , w s ) = ( p prior , 0 ) .
For this model, we use the NPDHG algorithm (18). Starting from l = 1 , we compute the pair of solutions ( p ( l ) , w ( l ) ) to the primal and dual problems (10) and (9) at hyperparameter t = t ( l ) using the previously computed pair of solutions ( p ( l 1 ) , w ( l 1 ) ) by setting the initial vectors z 0 = w 0 = w ( l 1 ) . For the stepsize parameters, we set θ 0 = 0 , τ 0 = 2 , and σ 0 = 1 / 2 A op 2 . We compute the update w k + 1 in (18) using the following proximal operator formula: for every group g { 1 . , 5 } , λ > 0 and w ^ R m ,
arg min w R m λ m g w 2 , g + 1 2 w w ^ 2 , g 2 = max 0 , 1 λ m g / w ^ 2 w ^ .
Finally, we let the NPDHG algorithms run for at least 40 iterations before checking for convergence. We stop the NPDHG algorithms when the optimality condition (23) is satisfied within some tolerance 10 5 :
max g { 1 , , 5 } E D ^ [ Φ ] E p k [ Φ ] 2 , g t ( 1 + 10 5 ) .

5.3.3. -Regularized MaxEnt Model

For this MaxEnt model, the smallest hyperparameter for which the solutions to the primal and dual problems (7) and (8) are equal to the prior distribution and zero is
t ( 0 ) = E D ^ [ Φ ] E p prior [ Φ ] 1 .
This follows from the optimality condition
E D ^ [ Φ ] E p s [ Φ ] t · ( w s )
and computing the smallest parameter t for which (24) is satisfied at ( p s , w s ) = ( p prior , 0 ) .
For this model, we use the NPDHG algorithm (18). Starting from l = 1 , we compute the pair of solutions ( p ( l ) , w ( l ) ) to the primal and dual problems (11) and (12) at hyperparameter t = t ( l ) using the previously computed pair of solutions ( p ( l 1 ) , w ( l 1 ) ) by setting the initial vectors z 0 = w 0 = w ( l 1 ) . For the stepsize parameters, we set θ 0 = 0 , τ 0 = 2 , and σ 0 = 1 / 2 A op 2 . Finally, we compute the update w k + 1 in (18) using Moreau’s decomposition ([78], Theorem 3.2.5): for every λ > 0 and w ^ R m ,
arg min w R m λ w + 1 2 w w ^ 2 2 = w ^ arg min w 1 λ 1 2 w w ^ 2 2 .
The second term on the right amounts to projecting w ^ on the 1 ball of radius λ . There are fast algorithms for doing do; we use Algorithm 1 described in [79]. Finally, we let the NPDHG algorithms run for at least 40 iterations before checking for convergence. We stop the NPDHG algorithms when the optimality condition (24) is satisfied within some tolerance 10 5 :
E D ^ [ Φ ] E p k [ Φ ] 1 t ( 1 + 10 5 ) .

5.4. Comparison of Timings

In this section, we compare the run times of our NPDHG algorithms with two state-of-the-art optimization algorithms for solving non-smooth MaxEnt models: the forward–backward splitting algorithm (specifically, Algorithm 5 in [65]; see also [40,66]) and the STRUCTMAXENT2 coordinate descent algorithm from [41,64]. All numerical experiments were performed on a single core Intel(R) Core(TM) i7-10750H CPU @ 2.60 GHz.
We initially chose the GLMNET implementation [26,33] in MATLAB 2023b over STRUCTMAXENT2 for the numerical comparisons, but we found that GLMNET produced unreliable numerical solutions when compared to both the NPDHG and forward–backward splitting algorithms. We also tried using GLMNET’s implementation in the R language, but to no avail. We think this problem arises because, as discussed in Section 3.2, the GLMNET algorithm approximates a MaxEnt model as a logistic regression model and then invokes a coordinate descent method tailored to logistic regression to approximate the solution to the MaxEnt model. Our observations suggest this approach does not work well for our data set. In contrast, we found that the STRUCTMAXENT2 algorithm produced correct numerical results.
For the forward–backward splitting algorithm, the stepsize parameters were set to 1 / A 2 and 1 (corresponding to τ and t 0 in Algorithm 5 of [65]) and, for computing the pair of solutions ( p ( l ) , w ( l ) ) at hyperparameter t ( l ) , the initial iterate was set to w ( l 1 ) . In addition, for the elastic net MaxEnt models, the acceleration quantity q = ( 1 α ) t ( l ) / ( A 2 + ( 1 α ) t ( l ) ) was employed. We used the stopping criteria of the NPDHG algorithms for the forward–backward splitting algorithms. For the coordinate descent algorithm, we modified the STRUCTMAXENT2 algorithm from ([41], pp. 305–306) to make it applicable to the elastic net penalty. For computing the pair of solutions ( p ( l ) , w ( l ) ) at hyperparameter t ( l ) , the initial iterate was set to w ( l 1 ) . We did not use the STRUCTMAXENT2 algorithm for the non-overlapping group lasso or -regularized MaxEnt models, as it was not designed for these MaxEnt models. We used the stopping criteria of the NPDHG algorithms for the STRUCTMAXENT2 algorithm.
Table 1 shows the average timings for computing the entire regularization path of the WUMI wildfire data set using the coordinate descent, forward–backward splitting, and NPDHG algorithms. All timings were averaged over five runs and, for the forward–backward splitting and NPDHG algorithms, they include the time required to compute all the stepsize parameters. All algorithms were implemented in MATLAB.
The NPDHG algorithm outperformed both the forward–backward splitting and coordinate descent algorithms by at least one order of magnitude. In particular, we observed that the NPDHG algorithm required far fewer iterations to achieve convergence compared to both the forward–backward splitting and STRUCTMAXENT2 algorithms. This difference is because the stepsize parameters for the NPDHG algorithm were much larger compared to either the forward–backward splitting or the STRUCTMAXENT2 algorithm. Indeed, the stepsize parameters for the NPDHG and forward–backward splitting algorithms are inversely proportional to the norms A op and A 2 , and for the wildfire data set these were A op 3.30 and A 2 854.08 . Thus, larger stepsize parameters were permitted thanks to the Kullback–Leibler divergence term in the NPDHG algorithm, enabling a major speedup gain.

5.5. Analysis of the MaxEnt Regularization Paths and Estimated Fire Probabilities

As a final validation step for the NPDHG algorithm, we use the fitted MaxEnt models to compute the normalized mean fire probability in each grid cell for all calendar months between 1984 and 2020. The probabilities are visualized in Figure 2 and Figure 3. In each case, we have chosen the prior distribution as a benchmark for our spatial plots of fire probability.
The spatial fire probabilities for α = { 0.95 , 0.40 , 0.15 , 0.05 } are shown in Figure 2. The range of α values roughly corresponds to varying the regularization from a purely l 1 norm ( α = 1 ) to a purely l 2 norm ( α = 0 ). For each value of α , we also consider the evolution of the spatial fire probability as we vary the hyperparameter t along the regularization path, or equivalently include additional features while fitting the MaxEnt model to wildfire data. Broadly, we observe that, for a fixed value of t / t max , the ratio of predicted to prior fire probability decreases in sharpness as α decreases. On the other hand, for a fixed α , decreasing t / t max values enables the model to make sharper distinctions between grid cells with high and low fire probability as evidenced by the sharper contrast between the prior and predicted fire probabilities. We also note a similar pattern in Figure 3 for the MaxEnt models with non-overlapping group lasso (corresponding roughly to the elastic net case with α = 1 ) and the l -MaxEnt models, all of which converge to the empirical distribution quicker than any of the elastic net cases.
In Figure 4, we show the cumulative number of non-zero coefficients at fixed intervals along the regularization path for different α values. The plot helps in visualizing the t / t max values at which new features are introduced in the elastic net MaxEnt model, with the dashed vertical lines indicating the first time a feature from a new group is selected. Across all α values, we find that features appear in the same order, with fire weather features being selected first, followed by topography, vegetation, human, and antecedent features. We tabulate the final set of non-zero features at the end of the regularization path for α = { 0.95 , 0.05 } across various groups in Table 2. These selected features are in good agreement with physical models of wildfire occurrence [42,43] as well as previous statistical analyses of wildfire drivers [29,30,31,44].

6. Discussion

In this paper, we have introduced novel first-order NPDHG algorithms that overcome the shortcomings of state-of-the-art algorithms for training large-scale, non-smooth MaxEnt models. The crucial ingredient common to our algorithms is the Kullback–Leibler divergence. Using it over the classical proximal operator makes it possible to train large-scale and non-smooth MaxEnt models much more efficiently than the state of the art. In particular, all stepsize parameters and iterations in our algorithms can be calculated on the order O ( m n ) operations, improving on the complexity bound of O ( min ( m 2 n , m n 2 ) ) operations for computing the optimal stepsize parameters of classical first-order optimization methods, such as the linear PDHG or forward–backward splitting methods. As a consequence, for a given tolerance level ϵ > 0 , our algorithms provably compute solutions using on the order of O ( m n / ϵ ) or O ( m n / log ( 1 / ϵ ) ) operations, the order depending on the smoothness of the MaxEnt model and which is optimal with respect to the Nesterov class of optimal first-order methods [39]. Moreover, the computational bottleneck consists of matrix–vector multiplications, which can be trivially parallelized, and so our algorithms exhibit scalable parallelism.
Finally, we have shown that the strong convexity of the Kullback–Leibler divergence with respect to the 1 norm allows for significantly larger stepsize parameters, thereby speeding up the convergence rate of our algorithms. This was, in particular, observed in Section 5, when we applied our algorithms to fit the WUMI wildfire data set [45] on several non-smooth MaxEnt models to estimate the probabilities of fire occurrences as a function of ecological features. Our algorithms outperformed the state-of-the-art forward–backward splitting and coordinate descent STRUCTMAXENT2 algorithms by at least one order of magnitude. They also yielded results that are in good agreement with physical models of wildfire occurrence [42,43] as well as previous statistical analyses of wildfire drivers [29,30,31,44]. Future work will explore the scalability of our algorithms for modeling daily scale wildfire probability [80]. It will also explore how the choices of different informative priors affect the MaxEnt models using prediction accuracy as a metric.
We expect our algorithms to provide efficient methods for solving non-smooth MaxEnt models that arise in large-scale machine learning applications beyond the wildfire application explored in this paper. An interesting future direction, albeit tangential to the focus of this work, would be to prove learning guarantees for the MaxEnt problem (3) for arbitrary choices of potential functions, similarly to [19,64], for the choice of 1 and 2 norms as potential functions. This would provide some means to assess the “quality” of the probability distribution estimated from the MaxEnt problem. Another direction that would be interesting and impactful is to extend our algorithms to continuous regularized MaxEnt models and prove rigorously that they work: interesting because the continuous version of the MaxEnt problem is an infinite dimensional problem, which makes this problem more technically challenging, impactful because such an algorithm would enable a much broader class of probability distributions to be used in MaxEnt modeling.

Author Contributions

Conceptualization, G.P.L., J.B. and J.D.; data curation, J.B.; formal analysis, G.P.L.; methodology, G.P.L. and J.B.; software, G.P.L. and J.B.; visualization G.P.L. and J.B.; writing—original draft G.P.L., J.B. and J.D.; writing—review & drafting G.P.L., J.B. and J.D. All authors have read and agreed to the published version of the manuscript.

Funding

G.P.L. and J.D. are supported by ONR N00014-22-1-2667. J.D. is also supported by DOE-MMICS SEA-CROGS DE-SC0023191. J.B. is supported by the Zegar Family Foundation and the National Science Foundation LEAP Science and Technology Center Award #2019625.

Institutional Review Board Statement

Not Applicable.

Data Availability Statement

The data that support the findings of this study are openly available in Zenoto at https://zenodo.org/records/7277980 (accessed on 1 April 2024). The code used to produce the results from this project can be found on Github at https://github.com/Gabriel-P-Langlois/maxent-paper-wildfire.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Appendix A. Derivation of Update (17)

We divide the derivation into two parts, first deriving an expression for the update p k + 1 in terms of the previous iterate p k and then deriving the explicit update for p k + 1 as in the third line of (16).
Part 1. First, note the minimization problem
min p Δ n τ k D KL ( p p prior ) τ k w k + θ k ( w k w k 1 ) , E p [ Φ ] + D KL ( p p k )
has a unique global minimum, denoted by p k + 1 , in the relative interior of the unit simplex Δ n for every k N . See ([38], Proposition 4.1) for details and note that Proposition 4.1 applies because the Kullback–Leibler divergence is strongly convex with respect to the 1 norm in its first argument.
Having established that the minimum p k + 1 exists and is unique, we introduce the Karush–Kuhn–Tucker multipliers ξ R and μ R n to write the minimization problem for p k + 1 in terms of an unconstrained optimization problem:
min p R n ξ R μ R n τ k D KL ( p p prior ) τ k w k + θ k ( w k w k 1 ) , E p [ Φ ] + D KL ( p p k ) + ξ 1 j = 1 n p ( j ) μ , p .
We can invoke the Karush–Kuhn–Tucker Theorem and use the linearity constraint qualifications to find that p k + 1 satisfies the first-order optimality condition
τ k log ( p k + 1 ( j ) / p prior ( j ) ) + τ k τ k w k + θ k ( w k w k 1 ) , Φ ( j ) + log ( p k + 1 ( j ) / p k ( j ) ) + 1 ξ [ μ ] j = 0
for each j { 1 , , n } [81]. Moreover, the complementary slackness condition μ , p k + 1 = 0 with the fact that p k + 1 > 0 for every j { 1 , , m } implies μ = 0 .
Next, we use the first-order optimality condition to compute p k + 1 ( j ) explicitly, rearranging (A1) in terms of log ( p k + 1 ( j ) ) , using μ = 0 , and taking the exponential yield
p k + 1 ( j ) = e ( ξ 1 ) / ( 1 + τ k ) 1 ( p prior ( j ) ) τ k p k ( j ) e τ k w k + θ k ( w k w k 1 ) , Φ ( j ) 1 / ( 1 + τ k )
Since j = 1 n p k + 1 ( j ) = 1 , we find that ξ satisfies the relation
e τ k ξ / ( 1 + τ k ) 1 = 1 j = 1 n ( p prior ( j ) ) τ k p k ( j ) e τ k w k + θ k ( w k w k 1 ) , Φ ( j ) 1 / ( 1 + τ k ) .
Hence,
p k + 1 ( j ) = ( p prior ( j ) ) τ k p k ( j ) e τ k w k + θ k ( w k w k 1 ) , Φ ( j ) 1 / ( 1 + τ k ) j = 1 n ( p prior ) τ k ( j ) p k ( j ) e τ k w k + θ k ( w k w k 1 ) , Φ ( j ) 1 / ( 1 + τ k )
for each j { 1 , , n } . This yields an explicit update for p k + 1 in terms of the previous iterate p k .
Part 2. We now induce on k to prove that the explicit expression for the update p k + 1 in the third line of (16) is correct. For k = 0 , we have
p 0 ( j ) = p prior ( j ) e z 0 , Φ ( j ) j = 1 n p prior ( j ) e z 0 , Φ ( j )
and so
p prior ( j ) τ 0 p 0 ( j ) e τ 0 w 0 + θ 0 ( w 0 w 1 ) , Φ ( j ) 1 / ( 1 + τ 0 ) = p prior ( j ) e ( z 0 + τ 0 ( w 0 + θ 0 ( w 0 w 1 ) ) ) / ( 1 + τ 0 ) , Φ ( j ) j = 1 n p prior ( j ) e z 0 , Φ ( j ) 1 / ( 1 + τ 0 ) .
Writing
z 1 = ( z 0 + τ 0 ( w 0 + θ 0 ( w 0 w 1 ) ) ) / ( 1 + τ 0 ) ,
the term inside the exponential on the numerator simplifies to z 1 , Φ ( j ) . Hence,
p prior ( j ) τ 0 p 0 ( j ) e τ 0 w 0 + θ 0 ( w 0 w 1 ) , Φ ( j ) 1 / ( 1 + τ 0 ) = p prior ( j ) e z 1 , Φ ( j ) j = 1 n p prior ( j ) e z 0 , Φ ( j ) 1 / ( 1 + τ 0 ) .
It follows on substitution that
p 1 ( j ) = p prior ( j ) e z 1 , Φ ( j ) j = 1 n p prior ( j ) e z 1 , Φ ( j )
for every j { 1 , , n } .
For arbitrary k N , we use the induction hypothesis
p k ( j ) = p prior ( j ) e z k , Φ ( j ) j = 1 n p prior ( j ) e z k , Φ ( j )
to find
p prior ( j ) τ k p k ( j ) e τ k w k + θ k ( w k w k 1 ) , Φ ( j ) 1 / ( 1 + τ k ) = p prior ( j ) e ( z k + τ k ( w k + θ k ( w k w k 1 ) ) ) / ( 1 + τ k ) , Φ ( j ) j = 1 n p prior ( j ) e z k , Φ ( j ) 1 / ( 1 + τ k ) .
Writing
z k + 1 = ( z k + τ k ( w k + θ k ( w k w k 1 ) ) ) / ( 1 + τ k ) ,
the term inside the exponential on the numerator simplifies to z k + 1 , Φ ( j ) . Hence,
p prior ( j ) τ k p k ( j ) e τ 0 k w k + θ k ( w k w k 1 ) , Φ ( j ) 1 / ( 1 + τ k ) = p prior ( j ) e z k + 1 , Φ ( j ) j = 1 n p prior ( j ) e z k , Φ ( j ) 1 / ( 1 + τ k ) .
It follows on substitution that
p k + 1 ( j ) = p prior ( j ) e z k + 1 , Φ ( j ) j = 1 n p prior ( j ) e z k + 1 , Φ ( j )
for every j { 1 , , n } , which proves the desired result.

Appendix B. Summary of Features Extracted from the WUMI Wildfire Data Set

We provide here a summary of all the features used in our statistical analysis. The features are aggregated to a 12 km spatial resolution during data preprocessing. Considering each feature’s M antecedent month average and maximum X-day running average components as distinct features, the total number of features considered in our analysis adds up to 35. We organize the features into five groups with each group reflecting a major environmental driver of fire occurrences. These groups are also an important ingredient of the non-overlapping group lasso MaxEnt model.
Table A1. Summary table for all input features organized by group, identifier, description, spatial resolution of raw data, and source of data.
Table A1. Summary table for all input features organized by group, identifier, description, spatial resolution of raw data, and source of data.
Feature GroupIdentifierDescriptionResolutionSource
Fire weatherVPDMean vapor pressure deficit5 kmClimgrid [82], PRISM [83]
VPD max X Maximum X-day running average of VPD; X { 3 , 7 } 9 kmUCLA-ERA5 [84]
TmaxDaily maximum temperature5 kmClimgrid
Tmax max X Maximum X-day running average of Tmax9 kmUCLA-ERA5
TminDaily minimum temperature5 kmClimgrid
Tmin max X Maximum X-day running average of Tmin9 kmUCLA-ERA5
PrecPrecipitation total5 kmClimgrid
SWE mean Mean snow water equivalent500 mNSIDC [85]
SWE max Daily maximum snow water equivalent500 mNSIDC
FM10001000 h dead fuel moisture4 kmgridMET [86]
FFWI max 7 Maximum 7-day running average of Fosberg Fire Weather Index9 kmUCLA-ERA5
WindMean wind speed9 kmUCLA-ERA5
LightningLightning strike density500 mNLDN [87,88]
Antecedent AntVPD M mon Average VPD in M antecedent months; M { 2 , 3 } 5 kmClimgrid
AntPrec M mon Average precipitation total in M antecedent months; M { 2 , 4 } 5 kmClimgrid
AntPrec lag 1 Mean annual precipitation in lag year 15 kmClimgrid
AntPrec lag 2 Mean annual precipitation in lag year 25 kmClimgrid
AvgSWE 3 mon Average snow water equivalent in 3 antecedent months500 mNSIDC
VegetationForestFraction of forest landcover30 mNLCD
GrasslandFraction of grassland cover30 mNLCD
ShrublandFraction of shrubland cover30 mNLCD
BiomassAboveground biomass map300 mRef. [89]
HumanUrbanFraction of land covered by urban areas30 mNLCD
Camp_numMean number of camp grounds1 kmOpen source
Camp_distMean distance from nearest camp ground1 kmOpen source
Road_distMean distance from nearest highway1 kmOpen source
PopdensityMean population density1 kmSILVIS [90]
HousedensityMean housing density1 kmSILVIS
TopographySlopeMean slope1 mUSGS
SouthnessMean south-facing degree of slope1 mUSGS

References

  1. Jaynes, E.T. Information theory and statistical mechanics. Phys. Rev. 1957, 106, 620. [Google Scholar] [CrossRef]
  2. Jaynes, E.T. Information theory and statistical mechanics. II. Phys. Rev. 1957, 108, 171. [Google Scholar] [CrossRef]
  3. Kapur, J.N. Maximum-Entropy Models in Science and Engineering; John Wiley & Sons: Hoboken, NJ, USA, 1989. [Google Scholar]
  4. Gu, Y.; McCallum, A.; Towsley, D. Detecting anomalies in network traffic using maximum entropy estimation. In Proceedings of the 5th ACM SIGCOMM Conference on Internet Measurement, Berkeley, CA, USA, 19–21 October 2005; p. 32. [Google Scholar]
  5. Bereziński, P.; Jasiul, B.; Szpyrka, M. An entropy-based network anomaly detection method. Entropy 2015, 17, 2367–2408. [Google Scholar] [CrossRef]
  6. Wainwright, M.J.; Jordan, M.I. Graphical Models, Exponential Families, and Variational Inference; Now Publishers Inc.: Hanover, MA, USA, 2008. [Google Scholar]
  7. Berger, A.; Pietra, S.A.D.; Pietra, V.J.D. A maximum entropy approach to natural language processing. Comput. Linguist. 1996, 22, 39–71. [Google Scholar]
  8. Chen, S.F.; Rosenfeld, R. A survey of smoothing techniques for me models. IEEE Trans. Speech Audio Process. 2000, 8, 37–50. [Google Scholar] [CrossRef]
  9. Pietra, S.D.; Pietra, V.D.; Lafferty, J. Inducing features of random fields. IEEE Trans. Pattern Anal. Mach. Intell. 1997, 19, 380–393. [Google Scholar] [CrossRef]
  10. Malouf, R. A comparison of algorithms for maximum entropy parameter estimation. In Proceedings of the COLING-02: The 6th Conference on Natural Language Learning 2002 (CoNLL-2002), Taipei, Taiwan, 31 August 2002. [Google Scholar]
  11. Ratnaparkhi, A. Maximum Entropy Models for Natural Language Processing; Springer: Boston, MA, USA, 2017; pp. 800–805. [Google Scholar]
  12. Tsujii, J.; Kazama, J. Evaluation and extension of maximum entropy models with inequality constraints. In Proceedings of the 2003 Conference on Empirical Methods in Natural Language Processing, Sapporo, Japan, 11–12 July 2003; pp. 137–144. [Google Scholar]
  13. Hayes, B.; Wilson, C. A maximum entropy model of phonotactics and phonotactic learning. Linguist. Inq. 2008, 39, 379–440. [Google Scholar] [CrossRef]
  14. Kwon, T.M. Tmc Traffic Data Automation for Mndot’s Traffic Monitoring Program; Technical Report; University of Minnesota: Twin Cities, MN, USA, 2004. [Google Scholar]
  15. Muttaqin, L.A.; Murti, S.H.; Susilo, B. Maxent (maximum entropy) model for predicting prehistoric cave sites in karst area of gunung sewu, gunung kidul, yogyakarta. In Sixth Geoinformation Science Symposium; SPIE-International Society for Optics and Photonics: Bellingham, WA, USA, 2019; Volume 11311, p. 113110B. [Google Scholar]
  16. Granot-Atedgi, E.; Tkačik, G.; Segev, R.; Schneidman, E. Stimulus-dependent maximum entropy models of neural population codes. PLoS Comput. Biol. 2013, 9, e1002922. [Google Scholar] [CrossRef]
  17. Tkačik, G.; Marre, O.; Mora, T.; Amodei, D.; Berry, M.J., II; Bialek, W. The simplest maximum entropy model for collective behavior in a neural network. J. Stat. Mech. Theory Exp. 2013, 2013, P03011. [Google Scholar] [CrossRef]
  18. Savin, C.; Tkačik, G. Maximum entropy models as a tool for building precise neural controls. Curr. Opin. Neurobiol. 2017, 46, 120–126. [Google Scholar] [CrossRef] [PubMed]
  19. Dudík, M.; Phillips, S.J.; Schapire, R.E. Performance guarantees for regularized maximum entropy density estimation. In Learning Theory; Shawe-Taylor, J., Singer, Y., Eds.; Springer: Berlin/Heidelberg, Germany, 2004; pp. 472–486. [Google Scholar]
  20. Dudík, M.; Phillips, S.J.; Schapire, R.E. Maximum entropy density estimation with generalized regularization and an application to species distribution modeling. J. Mach. Learn. Res. 2007, 8, 1217–1260. [Google Scholar]
  21. Elith, J.; Phillips, S.J.; Hastie, T.; Dudík, M.; Chee, Y.E.; Yates, C.J. A statistical explanation of maxent for ecologists. Divers. Distrib. 2011, 17, 43–57. [Google Scholar] [CrossRef]
  22. Kalinski, C.E. Building Better Species Distribution Models with Machine Learning: Assessing the Role of Covariate Scale and Tuning in Maxent Models. Ph.D. Thesis, University of Southern California, Los Angeles, CA, USA, 2019. [Google Scholar]
  23. Merow, C.; Smith, M.J.; Silander, J.A., Jr. A practical guide to maxent for modeling species’ distributions: What it does, and why inputs and settings matter. Ecography 2013, 36, 1058–1069. [Google Scholar] [CrossRef]
  24. Phillips, S.J.; Dudík, M.; Schapire, R.E. A maximum entropy approach to species distribution modeling. In Proceedings of the Twenty-First International Conference on Machine Learning, Banff, AB, Canada, 4–8 July 2004; p. 83. [Google Scholar]
  25. Phillips, S.J.; Anderson, R.P.; Schapire, R.E. Maximum entropy modeling of species geographic distributions. Ecol. Model. 2006, 190, 231–259. [Google Scholar] [CrossRef]
  26. Phillips, S.J.; Anderson, R.P.; Dudík, M.; Schapire, R.E.; Blair, M.E. Opening the black box: An open-source release of maxent. Ecography 2017, 40, 887–893. [Google Scholar] [CrossRef]
  27. Schnase, J.L.; Carroll, M.L.; Gill, R.L.; Tamkin, G.S.; Li, J.; Strong, S.L.; Maxwell, T.P.; Aronne, M.E.; Spradlin, C.S. Toward a monte carlo approach to selecting climate variables in maxent. PLoS ONE 2021, 16, e0237208. [Google Scholar] [CrossRef]
  28. Schnase, J.L.; Carroll, M.L. Automatic variable selection in ecological niche modeling: A case study using cassin’s sparrow (peucaea cassinii). PLoS ONE 2022, 17, e0257502. [Google Scholar] [CrossRef] [PubMed]
  29. Parisien, M.; Moritz, M.A. Environmental controls on the distribution of wildfire at multiple spatial scales. Ecol. Monogr. 2009, 79, 127–154. [Google Scholar] [CrossRef]
  30. Chen, B.; Jin, Y.; Scaduto, E.; Moritz, M.A.; Goulden, M.L.; Randerson, J.T. Climate, fuel, and land use shaped the spatial pattern of wildfire in California’s Sierra Nevada. J. Geophys. Res. Biogeosci. 2021, 126, e2020JG005786. [Google Scholar] [CrossRef]
  31. Yu, Y.; Saatchi, S.S.; Walters, B.F.; Ganguly, S.; Li, S.; Hagen, S.; Melendy, L.; Nemani, R.R.; Domke, G.M.; Woodall, C.W. Carbon Pools across Conus Using the Maxent Model, 2005, 2010, 2015, 2016, and 2017. 2021. Available online: https://daac.ornl.gov/CMS/guides/CMS_CONUS_Biomass.html (accessed on 1 April 2024).
  32. L’heureux, A.; Grolinger, K.; Elyamany, H.F.; Capretz, M.A.M. Machine learning with big data: Challenges and approaches. IEEE Access 2017, 5, 7776–7797. [Google Scholar] [CrossRef]
  33. Fithian, W.; Hastie, T. Finite-sample equivalence in statistical models for presence-only data. Ann. Appl. Stat. 2013, 7, 1917. [Google Scholar] [CrossRef] [PubMed]
  34. Darroch, J.N.; Ratcliff, D. Generalized iterative scaling for log-linear models. In The Annals of Mathematical Statistics; Institute of Mathematical Statistics: Waite Hill, OH, USA, 1972; pp. 1470–1480. [Google Scholar]
  35. Demchenko, Y.; Grosso, P.; Laat, C.D.; Membrey, P. Addressing big data issues in scientific data infrastructure. In Proceedings of the 2013 International Conference on Collaboration Technologies and Systems (CTS), San Diego, CA, USA, 20–24 May 2013; pp. 48–55. [Google Scholar]
  36. Thompson, N.C.; Greenewald, K.; Lee, K.; Manso, G.F. Deep learning’s diminishing returns: The cost of improvement is becoming unsustainable. IEEE Spectr. 2021, 58, 50–55. [Google Scholar] [CrossRef]
  37. Dyderski, M.K.; Paź, S.; Frelich, L.E.; Jagodziński, A.M. How much does climate change threaten european forest tree species distributions? Glob. Chang. Biol. 2018, 24, 1150–1163. [Google Scholar] [CrossRef] [PubMed]
  38. Darbon, J.; Langlois, G.P. Accelerated nonlinear primal-dual hybrid gradient algorithms with applications to machine learning. arXiv 2021, arXiv:2109.12222. [Google Scholar]
  39. Nesterov, Y. Introductory Lectures on Convex Optimization: A Basic Course; Kluwer Academic Publishers: Boston, MA, USA, 2004; Volume 87. [Google Scholar]
  40. Beck, A.; Teboulle, M. A fast iterative shrinkage-thresholding algorithm for linear inverse problems. SIAM J. Imaging Sci. 2009, 2, 183–202. [Google Scholar] [CrossRef]
  41. Mohri, M.; Rostamizadeh, A.; Talwalkar, A. Foundations of Machine Learning, 2nd ed.; MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  42. Parisien, M.; Dawe, D.A.; Miller, C.; Stockdale, C.A.; Armitage, O.B. Applications of simulation-based burn probability modelling: A review. Int. J. Wildland Fire 2019, 28, 913. [Google Scholar] [CrossRef]
  43. Mezuman, K.; Tsigaridis, K.; Faluvegi, G.; Bauer, S.E. The interactive global fire module pyre (v1.0). Geosci. Model Dev. 2020, 13, 3091–3118. [Google Scholar] [CrossRef]
  44. Buch, J.; Williams, A.P.; Juang, C.S.; Hansen, W.D.; Gentine, P. Smlfire1.0: A stochastic machine learning (sml) model for wildfire activity in the western united states. Geosci. Model Dev. 2023, 16, 3407–3433. [Google Scholar] [CrossRef]
  45. Juang, C.S.; Williams, A.P.; Abatzoglou, J.T.; Balch, J.K.; Hurteau, M.D.; Moritz, M.A. Rapid growth of large forest fires drives the exponential response of annual forest-fire area to aridity in the western united states. Geophys. Res. Lett. 2022, 49, e2021GL097131. [Google Scholar] [CrossRef]
  46. Ekeland, I.; Temam, R. Convex Analysis and Variational Problems; SIAM: New Delhi, India, 1999. [Google Scholar]
  47. Beck, A.; Teboulle, M. Mirror descent and nonlinear projected subgradient methods for convex optimization. Oper. Res. Lett. 2003, 31, 167–175. [Google Scholar] [CrossRef]
  48. Csiszár, I. Information-type measures of difference of probability distributions and indirect observation. Stud. Sci. Math. Hung. 1967, 2, 229–318. [Google Scholar]
  49. Kemperman, J.H.B. On the optimum rate of transmitting information. In Probability and Information Theory; Springer: Berlin/Heidelberg, Germany, 1969; pp. 126–169. [Google Scholar]
  50. Kullback, S. A lower bound for discrimination information in terms of variation (corresp.). IEEE Trans. Inf. Theory 1967, 13, 126–127. [Google Scholar] [CrossRef]
  51. Pinsker, M.S. Information and Information Stability of Random Variables and Processes; Holden-Day: San Francisco, CA, USA, 1964. [Google Scholar]
  52. Altun, Y.; Smola, A. Unifying divergence minimization and statistical inference via convex duality. In International Conference on Computational Learning Theory; Springer: Berlin/Heidelberg, Germany, 2006; pp. 139–153. [Google Scholar]
  53. Dewaskar, M.; Tosh, C.; Knoblauch, J.; Dunson, D.B. Robustifying likelihoods by optimistically re-weighting data. arXiv 2023, arXiv:2303.10525. [Google Scholar]
  54. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Stat. Soc. Ser. (Stat. Methodol.) 2005, 67, 301–320. [Google Scholar] [CrossRef]
  55. Tay, J.K.; Narasimhan, B.; Hastie, T. Elastic net regularization paths for all generalized linear models. J. Stat. Softw. 2023, 106, 1. [Google Scholar] [CrossRef] [PubMed]
  56. Foucart, S.; Rauhut, H. Sparse Solutions of Underdetermined Systems; Springer: New York, NY, USA, 2013; pp. 41–59. [Google Scholar]
  57. Guide, M.E.; Jbilou, K.; Koukouvinos, C.; Lappa, A. Comparative study of l 1 regularized logistic regression methods for variable selection. In Communications in Statistics-Simulation and Computation; Taylor & Francis, Inc.: Philadelphia, PA, USA, 2020; pp. 1–16. [Google Scholar]
  58. Zanon, M.; Zambonin, G.; Susto, G.A.; McLoone, S. Sparse logistic regression: Comparison of regularization and bayesian implementations. Algorithms 2020, 13, 137. [Google Scholar] [CrossRef]
  59. Lee, J.D.; Recht, B.; Srebro, N.; Tropp, J.; Salakhutdinov, R.R. Practical large-scale optimization for max-norm regularization. In Advances in Neural Information Processing Systems; Curran Associates: Red Hook, NY, USA, 2010; Volume 23. [Google Scholar]
  60. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning, 2nd ed.; Data Mining, Inference, and Prediction; Springer Series in Statistics; Springer: New York, NY, USA, 2009. [Google Scholar]
  61. Friedman, J.; Hastie, T.; Höfling, H.; Tibshirani, R. Pathwise coordinate optimization. Ann. Appl. Stat. 2007, 1, 302–332. [Google Scholar] [CrossRef]
  62. Friedman, J.; Hastie, T.; Tibshirani, R. Regularization paths for generalized linear models via coordinate descent. J. Stat. Softw. 2010, 33, 1. [Google Scholar] [CrossRef]
  63. Hastie, T.; Qian, J.; Tay, K. An Introduction to Glmnet. 2021. Available online: https://glmnet.stanford.edu/articles/glmnet.html (accessed on 17 February 2021).
  64. Cortes, C.; Kuznetsov, V.; Mohri, M.; Syed, U. Structural maxent models. In Proceedings of the 32nd International Conference on International Conference on Machine Learning—Volume 37, Lille, France, 6–11 July 2015; ICML’15. pp. 391–399. [Google Scholar]
  65. Chambolle, A.; Bock, T. An introduction to continuous optimization for imaging. Acta Numer. 2016, 25, 161–319. [Google Scholar] [CrossRef]
  66. Daubechies, I.; Defrise, M.; De Mol, C. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Commun. Pure Appl. Math. J. Issued Courant Inst. Math. Sci. 2004, 57, 1413–1457. [Google Scholar] [CrossRef]
  67. Dudík, M.; Schapire, R.E. Maximum entropy distribution estimation with generalized regularization. In International Conference on Computational Learning Theory; Springer: Berlin/Heidelberg, Germany, 2006; pp. 123–138. [Google Scholar]
  68. Mann, G.; McDonald, R.; Mohri, M.; Silberman, N.; Walker, D., IV. Efficient Large-Scale Distributed Training of Conditional Maximum Entropy Models. In Advances in Neural Information Processing Systems; Curran Associates: Red Hook, NY, USA, 2009; Volume 22. [Google Scholar]
  69. Andrew, G.; Gao, J. Scalable training of l 1-regularized log-linear models. In Proceedings of the 24th International Conference on Machine Learning, Vienna, Austria, 21–27 July 2007; pp. 33–40. [Google Scholar]
  70. Lee, S.-I.; Lee, H.; Abbeel, P.; Ng, A.Y. Efficient L1 regularized logistic regression. AAAI 2006, 6, 401–408. [Google Scholar]
  71. Yuan, G.-X.; Chang, K.-W.; Hsieh, C.-J.; Lin, C.-J. A comparison of optimization methods and software for large-scale l1-regularized linear classification. J. Mach. Learn. Res. 2010, 11, 3183–3234. [Google Scholar]
  72. Yu, H.-F.; Huang, F.-L.; Lin, C.-J. Dual coordinate descent methods for logistic regression and maximum entropy models. Mach. Learn. 2011, 85, 41–75. [Google Scholar] [CrossRef]
  73. Littell, J.S.; McKenzie, D.; Peterson, D.L.; Westerling, A.L. Climate and wildfire area burned in western u.s. ecoprovinces, 1916–2003. Ecol. Appl. 2009, 19, 1003–1021. [Google Scholar] [CrossRef]
  74. McElreath, R. Statistical Rethinking: A Bayesian Course with Examples in R and Stan; Chapman and Hall/CRC: Boca Raton, FL, USA, 2018. [Google Scholar]
  75. Gelman, A.; Carlin, J.B.; Stern, H.S.; Dunson, D.B.; Vehtari, A.; Rubin, D.B. Bayesian Data Analysis, 3rd ed.; Chapman & Hall/CRC Texts in Statistical Science; Taylor & Francis: Boca Raton, FL, USA, 2013. [Google Scholar]
  76. Figueiredo, M.A.T.; Nowak, R.D. Wavelet-based image estimation: An empirical Bayes approach using Jeffrey’s noninformative prior. IEEE Trans. Image Process. 2001, 10, 1322–1331. [Google Scholar] [CrossRef] [PubMed]
  77. Lions, P.-L.; Mercier, B. Splitting algorithms for the sum of two nonlinear operators. SIAM J. Numer. Anal. 1979, 16, 964–979. [Google Scholar] [CrossRef]
  78. Hiriart-Urruty, J.-B.; Lemaréchal, C. Convex Analysis and Minimization Algorithms I: Fundamentals; Springer Science & Business Media: Berlin/Heidelberg, Germany, 1993; Volume 305. [Google Scholar]
  79. Condat, L. Fast projection onto the simplex and the l 1 ball. Math. Program. 2016, 158, 575–585. [Google Scholar] [CrossRef]
  80. Keeping, T.; Harrison, S.P.; Prentice, I.C. Modelling the daily probability of wildfire occurrence in the contiguous United States. Environ. Res. Lett. 2024, 19, 024036. [Google Scholar] [CrossRef]
  81. Boyd, S.; Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  82. Vose, R.; Applequist, S.; Squires, M.; Durre, I.; Menne, M.; Williams, C.; Fenimore, C.; Gleason, K.; Arndt, D. Improved historical Temperature and Precipitation time series for U.S. climate divisions. J. Appl. Meteorol. Climatol. 2014, 53, 1232–1251. [Google Scholar] [CrossRef]
  83. Daly, C.; Gibson, W.; Doggett, M.; Smith, J.; Taylor, G. Up-to-date monthly climate maps for the conterminous United States. In Proceedings of the 14th AMS Conference on Applied Climatology, 84th AMS Annual Meeting Combined Preprints, Seattle, WA, USA, 11 January 2004. [Google Scholar]
  84. Rahimi, S.; Krantz, W.; Lin, Y.; Bass, B.; Goldenson, N.; Hall, A.; Jebo, Z.; Norris, J. Evaluation of a reanalysis-driven configuration of wrf4 over the western united states from 1980–2020. J. Geophys. Res. Atmos. 2022, 127, e2021JD035699. [Google Scholar] [CrossRef]
  85. Zeng, X.; Broxton, P.; Dawson, N. Snowpack change from 1982 to 2016 over conterminous United States. Geophys. Res. Lett. 2018, 45, 12940–12947. [Google Scholar] [CrossRef]
  86. Abatzoglou, J.T. Development of gridded surface meteorological data for ecological applications and modelling. Int. J. Climatol. 2013, 33, 121–131. [Google Scholar] [CrossRef]
  87. Wacker, R.S.; Orville, R.E. Changes in measured lightning flash count and return stroke peak current after the 1994 U.S. National Lightning Detection Network upgrade: 1. Observations. J. Geophys. Res. Atmos. 1999, 104, 2151–2157. [Google Scholar] [CrossRef]
  88. Orville, R.E.; Huffines, G.R. Cloud-to-Ground lightning in the United States: NLDN results in the first decade, 1989–98. Mon. Weather. Rev. 2001, 129, 1179–1193. [Google Scholar] [CrossRef]
  89. Spawn, S.A.; Sullivan, C.C.; Lark, T.J.; Gibbs, H.K. Harmonized global maps of above and belowground biomass carbon density in the year 2010. Sci. Data 2020, 7, 112. [Google Scholar] [CrossRef] [PubMed]
  90. Radeloff, V.C.; Hammer, R.B.; Stewart, S.I.; Fried, J.S.; Holcomb, S.S.; McKeefry, J.F. The Wildland-Urban Interface in the United States. Ecol. Appl. 2005, 15, 799–805. [Google Scholar] [CrossRef]
Figure 1. Wildfire activity in the western United States from 1984 to 2020. (Left) Fire locations of all fires (black dots) in the Western US MTBS-Interagency (WUMI) data set; also shown are three ecological divisions characterized by their primary vegetation type—forests (green), deserts (yellow), and plains (gray). (Right) Prior distribution indicating mean fire probability across all calendar months.
Figure 1. Wildfire activity in the western United States from 1984 to 2020. (Left) Fire locations of all fires (black dots) in the Western US MTBS-Interagency (WUMI) data set; also shown are three ecological divisions characterized by their primary vegetation type—forests (green), deserts (yellow), and plains (gray). (Right) Prior distribution indicating mean fire probability across all calendar months.
Entropy 26 00691 g001
Figure 2. Spatial probability plot for different hyperparameter values with elastic net penalty parameter α = { 0.95 , 0.40 , 0.15 , 0.05 } .
Figure 2. Spatial probability plot for different hyperparameter values with elastic net penalty parameter α = { 0.95 , 0.40 , 0.15 , 0.05 } .
Entropy 26 00691 g002
Figure 3. Same as Figure 2 but for (left) the non-overlapping group lasso with α = 1 and (right) the l MaxEnt models, respectively.
Figure 3. Same as Figure 2 but for (left) the non-overlapping group lasso with α = 1 and (right) the l MaxEnt models, respectively.
Entropy 26 00691 g003
Figure 4. Number of non-zero coefficients along the regularization path plots for elastic net penalty parameter α = { 0.95 , 0.40 , 0.15 , 0.05 } . The dashed vertical lines highlight the t / t max value at which the first feature of the group indicated by inset text is selected.
Figure 4. Number of non-zero coefficients along the regularization path plots for elastic net penalty parameter α = { 0.95 , 0.40 , 0.15 , 0.05 } . The dashed vertical lines highlight the t / t max value at which the first feature of the group indicated by inset text is selected.
Entropy 26 00691 g004
Table 1. Timings results (in seconds) for fitting the MaxEnt models described in Section 5.3. All timings are averaged over five runs.
Table 1. Timings results (in seconds) for fitting the MaxEnt models described in Section 5.3. All timings are averaged over five runs.
STRUCTMAXENT2Forward–Backward SplittingNPDHG
Elastic net ( α = 0.95 )5562.194208.01365.55
Elastic net ( α = 0.40 )1018.731407.22113.53
Non-overlapping group lassoN/A3036.38278.14
-regularizationN/A2534.65289.98
Table 2. List of non-zero features at the end of the regularization path for two elastic net penalty parameters, organized by different feature groups. See Appendix B for additional description of the selected features.
Table 2. List of non-zero features at the end of the regularization path for two elastic net penalty parameters, organized by different feature groups. See Appendix B for additional description of the selected features.
Feature Group α = 0.95 α = 0.05
Fire weatherTmaxTmaxVPD
PrecPrecTmin
WindWind VPD max 3
FM1000FM1000 VPD max 7
Tmin max 3 Tmin max 3 Tmax max 3
LightningLightning Tmax max 7
Tmin max 7 SWE max
TopographySlopeSlopeSouthness
Southness
VegetationGrasslandBiomassForest
Shrub
HumanCamp_distCamp_distRoad_dist
Camp_numCamp_num
UrbanUrban
PopdensityPopdensity
Antecedent AvgPrec 2 mo AvgPrec 2 mo AntPrec lag 2
AvgVPD 3 mo AntPrec lag 1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Provencher Langlois, G.; Buch, J.; Darbon, J. Efficient First-Order Algorithms for Large-Scale, Non-Smooth Maximum Entropy Models with Application to Wildfire Science. Entropy 2024, 26, 691. https://doi.org/10.3390/e26080691

AMA Style

Provencher Langlois G, Buch J, Darbon J. Efficient First-Order Algorithms for Large-Scale, Non-Smooth Maximum Entropy Models with Application to Wildfire Science. Entropy. 2024; 26(8):691. https://doi.org/10.3390/e26080691

Chicago/Turabian Style

Provencher Langlois, Gabriel, Jatan Buch, and Jérôme Darbon. 2024. "Efficient First-Order Algorithms for Large-Scale, Non-Smooth Maximum Entropy Models with Application to Wildfire Science" Entropy 26, no. 8: 691. https://doi.org/10.3390/e26080691

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop