1. Introduction
Functional Data Analysis (FDA) deals with the analysis and theory of data that are in the form of functions, images and shapes, or more general objects [
1,
2]. The atom of functional data is a function, where for each subject in a random sample one or several functions are recorded. While the term “functional data analysis” was coined by [
1,
3], the history of this area is much older and dates back to Greven et al. (2020) [
4] and Rao et al. (1950) [
5]. There are several ways that the data can called functional. For instance, in archaeological studies the form of a three-dimensional image of each bone can be presented as functional data [
6].
FDA has many applications in areas such as biostatistics, environmental monitoring, finance, and image analysis. For example, in biostatistics, FDA can be used to study the patterns of gene expression over time or to model the progression of a disease. In finance, FDA can be used to analyze the dynamics of financial markets over time, and to develop models for predicting future stock prices or other financial variables [
7].
The functional data comes in many forms; however, their definition is a based on functions which are often smooth curves [
8]. The smoothness is a characteristic that filter the noises inside the raw data as efficiently as possible [
9]. They are several approaches for representing functional data in smoothing form: basis expansion, least-squares, and roughness penalty. One of basis expansions is based on B-splines. In this approach, both functional coefficient and functional predictor expand in the form of B-splines and thereafter the regularized estimate of functional coefficient will be applied [
10]. Müller (2005) [
11] expressed that the regression coefficient can be expressed in orthonormal basis determined by the eigenfunctions of the covariance of functional regressor. Goldmsmith [
12] in 2011, combined both approaches introduced by Ramsay [
10] and Müller [
11]. This method has been considered in this study for smoothness and it is called roughness penalty because it penalizes fits that are too rough [
13].
In many cases, data are observed longitudinally and collected over time. Longitudinal data is epansion of a time series, which is a sequence of data points that are ordered based on time, such that each data point corresponds to a specific moment in time [
14]. Longitudinal data analysis can provide more accurate estimates of the effects of interventions or treatments over time [
15]. The mixed-effects regression models have been used to study the longitudinal data [
16]. These models allow us to make subject-specific conclusions and present statements about individuals changes, trends, or effects over time.
Sometimes in practice, the data has both characteristics of functional and longitudinal. One of the case can be when the outcome is not functional; however, covariates are collected functionally and longitudinally over one-time scale. People often use two approaches. One is based on function models where functional covariates have been smoothed according to roughness penalty and it is called Longitudinal Penalized Functional Regression (LPFR) [
17]. The other one is built on according to principal component analysis for longitudinal functional data [
18]. One of the main disadvantage of both approaches is the functional covariates are remained constant over time. But sometime in practice, we might have situation where functional covariates and scalar outcome have been changed over time which is considered in this study. Longitudinal functional models have not been studied to incorporate longitudinal functional for multiple time scales; however; Kundu et al. (2016) [
19] published their work about the same data situation for one time-scale. Therefore, this study is an extension of Kund’s work to incorporate multilevel longitudinal functional covariates.
Studying two time scales is important in a wide range of fields, including physics, chemistry, biology, engineering, economics, and more. It allows researchers to gain a better understanding of complex systems and processes that operate over different time frames [
20].
One of the key benefits of studying two time scales is that it can help researchers identify the underlying mechanisms that contribute to the behavior of a system. For example, in physics, studying the behavior of particles over different time scales can help researchers understand how those particles interact with each other and how their interactions lead to the formation of larger structures.
In chemistry, studying chemical reactions over different time scales can help researchers identify the intermediates and transition states that are involved in the reaction, and help them design more efficient and effective reactions [
21].
Studying two time scales can also help researchers design and optimize systems for specific applications. For example, in engineering, studying the behavior of materials over different time scales can help researchers design materials with specific mechanical, electrical, or chemical properties.
Overall, studying two time scales is critical for advancing our understanding of complex systems and processes, and for developing new technologies and applications that can improve our lives.
Studying physical activities with two time scales is important because it allows us to understand how different processes operate at different timescales and how they interact with each other. Physical activities involve a wide range of processes that occur at different timescales, from short-term processes such as muscle activation and joint movements, to longer-term processes such as training adaptations and recovery [
22].
By studying physical activities with two time scales, researchers can better understand the underlying mechanisms that contribute to performance and health outcomes [
23]. For example, studying the short-term and long-term effects of exercise on muscle function and metabolism can help us understand how exercise training improves overall health and athletic performance [
24].
Furthermore, studying physical activities with two time scales can help inform the development of effective training programs and interventions. By understanding how different processes interact over time, we can design training programs that are optimized for different goals, such as improving strength or endurance, and minimizing the risk of injury or overtraining [
25].
Overall, studying physical activities with two time scales is critical for advancing our understanding of human performance and health and for developing effective strategies to optimize physical activity and exercise programs [
26].
The proposed model was appropriate to apply on the data collected by the Active Schools Institute of the University of Northern Colorado. They were interested in the associations between daily and weekly activity profile, as measured by accelerometers, and academic and behavioral outcomes. To address these interests, students from one primary school in a suburban area in the western United States wore accelerometers during the school day for 5 consecutive days over 4 different weeks of the year. Corresponding academic and behavioral data were also obtained.
The data collected from the Active Schools Institute assumed daily student activity crossed within weeks. Due to nature of this data, establishing a multilevel mixed-effects regression model, including demographics or teacher effects, and accelerometer wear-time was essential. Hence, the proposed model was suited to apply on the physical activity data set.
As a consequence, this article addresses the following questions:
According to
Section 2, the following question has been addressed: How can we represent a longitudinal functional regression model with time-varying regression functions for multiple time scales when time components are crossed?
Under
Section 2.3, the following question is addressed: How can the mean parameters of the propsed model be estimated?
A response to this question can be found in
Section 5: How can the model and estimation be implemented in software?
A detailed answer to this question can be found in
Section 3: How do the proposed model compare to similar existing models with single time scales in terms of model MSE, across sample sizes and levels of association?
2. Intensive Longitudinal Functional Model with Multiple Time Scales
The intensive longitudinal method represents sequences of repeated measurements frequently recorded to characterize a separate change process for each subject [
27].
Intensive longitudinal data refers to data that is collected repeatedly and frequently over a relatively short period of time, often at a high frequency (e.g., several times a day). This type of data is particularly useful for studying processes that unfold over time, such as mood fluctuations, stress reactivity, or medication effects [
27,
28].
Intensive longitudinal data can be collected using a variety of methods, including ecological momentary assessment (EMA), experience sampling methods (ESM), ambulatory monitoring, and diary or journal methods [
29]. These methods typically involve asking participants to report on their experiences, thoughts, and behaviors at multiple points throughout the day, either through self-reports on electronic devices (such as smartphones or smartwatches), or by filling out paper or online forms [
30].
The analysis of intensive longitudinal data requires specialized statistical methods that can capture the dynamic nature of the data, such as time-series analysis, multilevel modeling, and dynamic systems modeling. These methods allow researchers to examine how variables change over time, how they are related to each other, and how they are influenced by external factors [
31].
Intensive longitudinal data has been used in a variety of fields, including psychology, medicine, and public health, to study phenomena such as mood disorders, substance use, physical activity, and sleep patterns.
This study considers a longitudinal model where the time scales of functional predictors cross. Additionally, the proposed model is suitable when data is collected frequently over a short period of time, often at a high frequency (e.g., several times a day). As a consequence, the proposed model may be referred to as an Intensive Longitudinal Functional Model with Multiple Time Scales (ILFMM) [
32].
A specific estimation process is also included in the ILFMM. Estimating parameters is performed using a generalized ridge estimate. As a result of this estimation process, a linear mixed model representation can be derived. By using a mixed model approach, tuning parameters can be automatically selected. It is also possible to estimate parameters by fitting a linear mixed model based on this linear mixed model representation. It is worthwhile to note that the model presented has some advantages. There is no restriction on the time course of regression functions. In addition, it is possible to incorporate the structure of the regression function directly into the estimation process [
19]. The approach described by Kundu et al. (2016)[
19] was extended here to incorporate longitudinal functional predictors spanning multiple time scales.
2.1. Statistical Model of ILFMM
To represent a longitudinal functional regression model with time-varying regression functions for multiple time scales when time components are crossed, assume
presents univariate predictors,
denotes a predictor function for the
ith subject at timepoint
, and
denotes a predictor function for the
ith subject at timepoint
, where
,
, and longitudinal-time points
t can be decomposed in terms of
and
, where each
(
) corresponds to one value each from
and
. It is assumed that each observed predictor is sampled at the same
p locations,
, and there are an equal number of observations per subject (i.e.,
are equal). Let
and
be the
vector of values sampled from the realized functions
and
, respectively. Then, the observed data have a form of
where
is a scalar outcome,
is a
column vector of measurements on
K scalar predictors,
is the sampled predictor from the
ith subject corresponding to time
, and
is the sampled predictor from the
ith subject corresponding to time
. The longitudinal functional linear model with the scalar outcome and functional predictor along with multiple time scales could be written as
where
are functional coefficients at time
,
are functional coefficients at time
,
is the vector of
r random effects pertaining to subject
i, and
denote the subject-specific random effect and random error term, respectively.
Similar to a linear mixed model with time-related slope for longitudinal data, it is assumed that
and
can be decomposed into several time-invariant component functions
and
as follows:
where
and
are functions of
and
, respectively. Additionally,
and
are functions of
s. In general,
and
have three components:
,
, and
s. The time component
enters into
through
and the functional component enters into
through
. Additionally, the time component
enters into
through
and the functional component enters into
through
. The model is flexible with any function of
and
with
(e.g.,
or
and
or
). Additionally,
and
is distributed as
. It is assumed that
are independent,
are independent whenever
or all, and
and
are independent if
.
In Model (
1)
is the standard fixed effect. In addition,
is the standard multilevel random effect and
is the subject and time specific functional effect corresponding to level
w. Here
is the subject and time specific functional effect corresponding to level
d. And both the variances and covariances across time are assumed to be constant as
and
, where
and
. This implies a compound symmetry assumption for the variances and covariances.
Model (
1) can be written based on Equations (
2) and (
3) as
A primary goal in developing Model (
4) is the estimation of
,
,
, and
.
2.2. Matrix Form of ILFMM
In the following an explanation will be given demonstrating how the proposed model can be written in matrix form. For this purpose, the observed values and the design matrix of fixed effect of Model (
1) can be considered as:
It is important to point out that it is assumed there are an equal number of observations per subject (i.e.,
are equal). In addition, the matrices of functional coefficients,
and
, have the form
where
and
So Model (
4) would be
where
is the associated coefficient vector,
and
are coefficient vectors of function predictors,
is a vector of random effects and
is the corresponding design matrix.
In this framework, the estimation approach is an extension of generalized ridge regression that takes the form of a penalty to the longitudinal setting in a manner that allows the estimated regression functions and to vary with time.
2.3. Estimation Based on the Mixed Model Representation
The goal of this section is to estimate parameters based on the matrix form of Model (
4). This model was described in
Section 2.2 and it can be written as:
where
,
,
, and
are known matrices,
is an unobservable vector of fixed effects,
and
are unobservable functional vectors of fixed effects of two time scales, and
is an unobservable vector of random effects with
, and
. Let
and
, then
is distributed as
because
and
In addition, it is assumed that
and
are known.
The well-known mixed model equation, developed by Henderson (1950) [
33], can be applied at this time. Henderson’s mixed model equation is similar in spirit to normal equations, but the mixed model equation simultaneously provides Best Linear Unbiasted Estimators (BLUEs) and Best Linear Unbiased Predictors (BLUPs) by estimating the variance component
.
Theorem 1. If is a solution to the mixed modelwhere , , , and , then is a BLUE of and simultaneously is a BLUP of if and are known and also by estimating the variance component . Theorem 1 demonstrates how the mixed Model (
11) can provide BLUEs and BLUPs simultaneously.
Appendix A.1 includes the proof of the Theorem 1.
2.4. Roughness Penalties Estimate
In this section a roughness penalties estimate based on the matrix form of Model (
4) is discussed, which was described in
Section 2.2. As mentioned, the mixed model can be written as:
where
,
,
, and
are known matrices,
is an unobservable vector of fixed effects,
and
are unobservable functional vectors of fixed effects for two time scales, and
is an unobservable vector of random effects with
, and
. Let
and
, then
is distributed as
(see
Section 2.3).
It is also assumed that is known.
Theorem 2. For each and , let and be penalty operators for and with and as the associated tuning parameters. A generalized ridge estimate associated with Model (13) for minimizing β, γ, and ηis obtained aswhereand In Theorem 2, an expression for the generalized ridge estimate of Model (
13) will be derived. The results of both theorems illustrated that the ridge estimate is BLUP from an equivalent mixed model; hence, the estimation procedure takes advantage of the equivalence between penalized least-squares estimation and linear mixed model representation. An
Appendix A.2 contains the proof of the Theorem 2.
2.5. Selection of Time-Structure in and
In the proposed approach there is freedom to choose a time-structure for both and . The matter of selecting appropriate time-structure in and are analogous, in essence, to that of selecting time structure in a linear mixed-effects model.
For example, selection of time-structure or or is similar to selection of variables in mutiple linear regression models based on linear mixed-effects model.
When comparing different time structures, it is common to apply Akaike Information Criterion (AIC). The formula , where l refers to a Restricted, or a Residual Maximum Likelihood (REML) estimation of mixed models and p refers to the number of parameters of the model, for information criterion (AIC) was used to determine time-structure in and .
2.6. Selection of a, b, c, and d for Decomposition Penalties
The set of values for each of a, b, c, and d are considered in this study. Those are treated as weights of a trade-off between preferred and non-preferred subspaces with constant and constant.
Additionally, REML is used to estimate tuning parameters when b and d are considered as fixed values and a grid search determines a and c values. These values are used to jointly select the tuning parameters.
Then, the selected values of a and c can be applied to the preferred subspace. The formula for information criterion (AIC) creates values for a and c, where l refers to a Restricted, or a Residual Maximum Likelihood (REML) estimation of mixed models and p refers to the number of parameters of the model.
3. Simulation Steps
In this section, the process of simulating the proposed model is explored. Without loss of generality, it was assumed that
and
; however,
a and
c could be varied on the set values of
. The largest values of
a and
c indicate there was much prior information in the estimation process. The following was used to generate the response data for 100 subjects by simulating functional covariates
in 4 time points
and functional covariates
in 5 time points
:
where
,
. Functional covariates
and
have the following form:
and
. Consequently, a rewritten of Model (
16) would be
Functional covariates
were generated by considering three sets of sampling points: wide, moderate, and narrow corresponding to one, five, and two bumps, respectively [
19].
As an example, the first set of 100 columns in was represented as follows:
- a.
We have recorded all of the observed values for the first subject at the first time point in the first row of the matrix.
- b.
As a result, all observation values for at all time points and all subjects in each following row have been arranged. This is so that each row represents an observation value for from a specific time point and subject. In accordance with , the rows were ordered by time and subject.
Having constructed the first set of 100 columns, we next proceeded to construct the second set of 100 columns by multiplying the first 100 columns by .
Since it was impossible to generate the curves with functional form only by considering discrete observations at different points of the domain interval, it was necessary to consider some set of sampling points on the interval
[
19]. To simulate the true functional forms of the curves, prior knowledge about the functional variable was applied. Due to this, the observed value of
was calculated considering the degrees of curvature at the wide, moderate, and narrow bumps [
19]. White noise was added to the predictor functions to account for the instrumental measurement noise [
19].
Functional covariates were generated by considering three sets of sampling points: wide, moderate, and narrow corresponding to one, five, and two bumps respectively as and The first set of 100 columns in were represented similar to process of presenting as discussed ealier. Consequently, the observed value of was generated in the same way as .
The functional coefficients
,
were generated according to [
19] considering different degrees of curvature at the wide, moderate, and narrow bumps. But
, and
were generated as follows with bumps centered at
, and
,
where
,
,
, and
correspond to amplitude and degree of curvature respectively. In fact, “bump” is the same as the “turning” of the polynomial graph. While the proper term is local maximum, the term “bump” referred to the “turning” of functional coefficients. The amplitude is the height from the centerline to the peak (dip) of the polynomial, and it measures the height from highest to lowest points divided by two. The amplitude values as
and
were considered to simulate functional coefficients [
19]. Another term that played an important role in simulating was degree of curvature. It was referred to the degree of a polynomial [
34]. The amplitude and curvature used to generate the functional coefficients
, and
are specified in
Table 1.
The model was applied to analyze the physical activities of kids with logitudonal functional data with two-time scale. The results suggested that
and
were reasonable choices for fixed parameters in the simulation study. Having made these choices the fixed effect, the random effect, and errors were generated and are shown in
Table 2.
Throughout this simulation the following decomposition penalties, an extension of Kundu et al. (2016) [
19] for multiple time scales, were considered:
and
while
a and
c were varied over
and 3, also structured penalties were
and
R 3.5.2 GUI 1.70 El Capitan build (7612) and RStudio Version 1.1.383 were used to analyze the simulated data. The method discussed in this section was implemented in the R package
nlme via the
lme function. The R code has been uploaded to GitHub and it is viewable at
https://github.com/Mostafa2020-bit/PackageILFMM since November 2020.
About 100 data were generated for two different sample sizes(100 and 200) along with two different ( and ).
The primary interests were in the estimation of the functional coefficients and the measurement of the squared distance between the fitted and the true value of . Estimation error was summarized in terms of the Mean Squared Errors (MSE) of the functional coefficients. Also, the MSE of was obtained.
With respect to the MSE of both the functional covariates and the response, we investigated the effect of increasing the sample size and on these values of . If these were increased, would , , , , and decrease or not? In this scenario, the following criteria were computed: where is the estimate of and each observed predictor is sampled at the same p locations, . Similarly, , , were computed; however, the calculation of was different because of its discrete characteristics and it was computed as where denotes the model fitted values of the true .
The , , , and could provide empirical evidence on whether the estimates are close to the corresponding true function or not.
To determine the effect of the decomposed penalty weights a, b, c, and d without loss of generality, b and d were considered a fixed value of 1. Then a and c were increased up to 3 on an exponential scale (i.e., ). The reason for this was to determine which combination of a, b, c, and d values would improve the estimation of , , , and . In this manner, an evaluation of estimation performance would remain almost unchanged.
Decomposition penalty values were selected according to the formula for AIC. It was expected that minimized AIC for selected decomposition penalty values leads to a minimized MSE for different sample sizes and s. AIC values were computed for all combinations of a and c values. Consequently, the simulation parameters were a and c values, the four sample sizes, and the four s.
4. Simulation Conclusions
The simulation study illustrated the potential advantage of an Intensive Longitudinal Functional Model with Multiple time scales (ILFMM) estimate in exploiting an informed, structured penalty. The simulation suggested that as the sample size and
increased, the MSE for both functional coefficients and the fitted values decreased,
Figure 1 and
Table 3.
Furthermore, sample size had a larger impact for smaller
, and also
had a greater impact for smaller sample size,
Figure 2.
These results provided empirical evidence that the ILFMM estimates of functional coefficients were close to the true functional estimate (basically unchanged). In other words, the estimation of the functional coefficients rose to the level of the estimation of true functions. These results suggested that AIC could guide the choice of ridge weights,
Figure 3 and
Figure 4.
The results for AIC are displayed graphically in
Figure 3. It has been demonstrated, when c = 3, AIC is minimized among all cases. Therefore, the choice of decomposition penalties is a = 3, b = 1, c = 3, and d = 1. Also with sufficiently large values there was minimal impact on the estimation performance.
In addition, the model with the selected ridge penalties had the lowest MSE values of
as compared to the model with the defined selection of ridge penalties,
Figure 4 and
Table 3. These results implied that the model with the selected ridge penalties would perform better than the model with the basic selection of ridge penalties.
Furthermore, for smaller sample sizes and s, the ILFMM estimate may oversmooth the estimated regression function. However, by increasing the sample size to 200 or to , it was observed that the average ILFMM estimate of the functional coefficients approached the true functions.
All in all, the model with ridge penalties and appears to perform better than the remaining models with other selections of the ridge penalties.
In the absence of a model with two time scales for comparison, it was decided to evaluate the effect of sample size and on MSE of functional coefficients, the fitted and AIC values, which would indicate whether the patterns are consistent with those seen in previous models.
Kundu’s approach was based on a single time scale. The results of the proposed approach was lined up with consideration to a single time scale comparing AIC values [
19] (i.e., single versus multiple time scales).
Also, Kudu concluded the selected ridge penalties had the lowest MSE values as when compared to the model with the basic selection. This demonstrated that the selected ridge penalties were reasonable. In addition, when the sample size and
were increased, MSE values were decreased. Furthermore, sample size had a larger impact for smaller
, and also
had a greater impact for smaller sample size [
19].
5. Physical Activity Study Application
The proposed method could be implemented in a
ILFMM() (i.e.,
Intensive
Longitudinal
Functional
Model with
Multiple time scales) package in R. The author has written the
ILFMM() package for a longitudinal functional model with multiple time scales along with scalar outcome, multiple functional predictors, one or more scalar covariates, and subject-specific random intercepts through mixed model equivalence. This package also works for a longitudinal functional model with a single time scale. The
ILFMM() package has been uploaded to GitHub and it is viewable at
https://github.com/Mostafa2020-bit/PackageILFMM sicne Novebmer 2020.
It is more common in the Physical Activity (PA) field to calculate the amount of PA a child gets in a single day, rather than a week. So this might be why it has been difficult to find information on weekly amounts of PA. Consequently, it is standard in PA projects to get week data by taking the average of all weekdays [
35,
36].
Up until about five years ago most physical activity researchers only measured and reported five days of activity at a time. Most reviews of handling accelerometer data focus on the epochs, the device or the type of outcome reported (Moderate-to-Vigorous Physical Activity (MVPA) or light PA) [
37].
In PA field, the daily accelerometer-assessed time in MVPA over a week is usually calculated as the mean of data over 7 days. For participants with less than 7 valid days of data, the following formula is common to use for standardizing measurements to one week for all participants [
37]:
mean daily weekday MVPA time mean daily weekend MVPA time
The dataset was stored in a secure password-protected Super R machine server utilized by IT at University of Northern Colorado for the Active Schools Institute. The author had permission from the Active Schools Institute to access the dataset. The data used included the five-second records from all students in school 1 for all weeks 1, 2, 3, and 4 over five days of school time. The standard approach was used to get week data.
For all weeks the “Axis-1” measurements ranged approximately from 0 to , with a mean of , where “Axis-1” referred to the physical activities over x-axis coordinate.
For week 1, many observations were irrelevant (i.e., zeros recorded late in the evening), so observations were truncated to consider only the time between 8:00 AM and 3:30 PM. For example, by considering the first day of data collection the number of observations were . Further limitation on the number of observations came from the restriction of the hours from 8:00 AM to 3:30 PM on the first day. This resulted in a final data set of observations. This is around 5900 observations per student. This number is close but a little high for 5-second observations. It would be anticipated that there would be 5400 observations per student (12 per minute, 60 min. per hour, h). However, 8 students show twice as many observations, suggesting that they have two days labeled as 9 November 2016. For all other weeks there was no such a problem.
For all weeks, the Stata default for time-stamps was New Year’s 1960. All official time stamps showed dates around this time, but for analysis, dates were inherited from the “Date” variable that the research team included.
Time stamps were translated in an unexpected way. For Day 1, all activity appeared to begin at 1:00 AM for each student. We assumed this should have been 8:00 AM and translated the time-stamps accordingly. For all other days, recording appeared to begin at 5:00 PM, but all activity appeared to begin at midnight. We again assumed these times should have been 8:00 AM and translated time-stamps accordingly. We then removed all observations outside the hours of 8:00 AM through 3:30 PM. Plots of the five-second data are usually “smoothed” to show the true patterns in the data. Smoothing is a method of taking the data in “windows of time” to estimate means or regression trends. The size of the window has a strong impact on the graphic produced but is subjective, and it is usually determined by a span value.
In designing the matrix for weekly activities, 162,000 data points per subject were considered for each week. The researcher chose 162,000 data points because it was large enough to capture physical activity patterns from 8:00 AM to 3:30 PM. This implies 684,000 observations per subject for four weeks (i.e., 162,000 × 4 = 684,000). And since 80 subjects were considered, the total number of observations were 51,840,000. These design matrices are graphically shown for 11 subjects in
Figure 5 and
Figure 6. It provides us with information on the overall patterns of physical activity on a daily and weekly basis.
For daily physical activity analysis, we chose 162,000 because it was large enough to capture physical activity patterns every five seconds from 8:00 AM to 3:30 PM. This implies 3,240,000 observations per subject for five days across four weeks (i.e., 162,000 × 5 × 4 = 3,240,000). And since 80 subjects were considered the total number of observations were 259,200,000. These design matrices are graphically shown in
Figure 7 and
Figure 8.
The proposed model with two-time scales along with
,
,
, and
as ridge penalties was applied to real data as
where
,
,
, and
. The expanded form of Model (
19) would be
The ILFMM estimates were obtained as BLUP assuming and the subject-specific random intercepts . The model was fitted into ILFMM() R package and the estimates of and were and , respectively.
We compared models based on AIC values with a different selection of weight penalties and time structures (linear, quadratic, and cubic). The results are summarized in
Table 4.
Based on the AIC, Model 16, with a cubic time structure for days and linear time structure for weeks and weight penalties , , , and appears to be better than the other models.
6. Discussion
The proposed intensive longitudinal functional model with multiple time-varying scales with scalar outcome, multiple functional predictors, one or more scalar covariates, and subject-specific random intercepts through mixed model equivalence has been defined. There were three primary advantages of this framework. First, this model estimated a time-dependent regression function. Second, it was also able to incorporate structural information into the estimation process. And third, it was easily implemented through linear mixed model equivalence.
One of the significant limitations of this study was the hardware needs for the organization and analysis of the data. Storage space to house the data, networking bandwidth to transfer it to and from analytics systems, and computing resources to perform those analytics were very time-consuming.
The physical activity data study results indicate that the proposed method is an appropriate one for the tested application. By using the proposed model, we could investigate what kind of time-structure for activity patterns would adequately describe the relationship between the subjects’ daily total magnitude and their weekly physical activities. Likewise, it described this relationship for daily physical activities.
So, the interest of the Active Schools Institute has been addressed through the proposed model. The proposed model concluded that the time structure of weekly activity was linear, but for daily activity was cubic. This conclusion was reasonable because the general patterns of movement throughout the day had more fluctuations than weekly movement. The weekly movements showed more smoothed patterns. Also, several dips in data points around the sedentary category for daily movement were observed. However, for weekly movement, it appeared this sedentary behavior was changed (i.e., it is needed to consider two separate definitions for sedentary behavior for weekly and daily activities). This analysis also gives us a slightly better sense of how movement intensity changes over time for both time scales.
The following are recommendations for future studies. One area of further investigation is the effect of different time structures. In this study, a linear form for simulation was considered; however, other forms of , such as or could be explored. A relative improvement in AIC due to the use of these complex structures would be anticipated. For instance, it would be expected to have a minimal AIC associated with in comparison to AIC for .
In the proposed model, the ridge weights were set as while a and c varied on an exponential scale. Larger values of a and c indicated greater emphasis of prior information on the estimation process. Future study could investigate other scales of ridge weights to determine more accurately and quickly the estimation process.
A possible extension of this work could be to incorporate multiple functional predictors. For example, in PA projects, the physical activity and patterns might be studied over an entire year. This implies there would be three-time scales: days, weeks, and years. Consequently, an extension of the proposed approach would be needed.
Additionally, an interesting and powerful use of the framework underlying the proposed model is when the response variable is binary. In this case when the response variable is binary, different problems arise. Multicollinearity and high dimensionality prejudice the estimation of the model and the interpretation of its parameters. This prejudice can be overcome by using an intensive longitudinal multilevel functional logistic regression model with principal component analysis. This study could be extended to any differentiable Hilbert space for analyzing of image, spatial, and spatial-temporal data.