Abstract
Various dimensionality reduction (DR) schemes have been developed for projecting high-dimensional data into low-dimensional representation. The existing schemes usually preserve either only the global structure or local structure of the original data, but not both. To resolve this issue, a scheme called sparse locality for principal component analysis (SLPCA) is proposed. In order to effectively consider the trade-off between the complexity and efficiency, a robust L2,p-norm-based principal component analysis (R2P-PCA) is introduced for global DR, while sparse representation-based locality preserving projection (SR-LPP) is used for local DR. Sparse representation is also employed to construct the weighted matrix of the samples. Being parameter-free, this allows the construction of an intrinsic graph more robust against the noise. In addition, simultaneous learning of projection matrix and sparse similarity matrix is possible. Experimental results demonstrate that the proposed scheme consistently outperforms the existing schemes in terms of clustering accuracy and data reconstruction error.
1. Introduction
In recent years, the development of high-throughput data processing schemes in diverse fields including pattern recognition, data mining, and computer vision has resulted in an exponential growth in the amount of harvested data with respect to both the dimensionality and size. However, large amount of redundancy and noise in the data cause significant spatial instability, computational complexity, and unfavorable representation. In relation to this problem, dimensionality reduction (DR) has been identified as an effective approach due to its capacity in dealing with a large amount of data and potential for overcoming what is called the “curse of dimensionality” [1]. It also offers a greater scope of model generalization and accomplishes the tasks with a high degree of computational efficiency.
To date, a variety of DR techniques have been developed for projecting the original data of high-dimensional space into a lower dimensional subspace. They can be classified in two ways: global dimensionality reduction (GDR) and local dimensionality reduction (LDR) [2]. The GDR techniques assume that all pairwise distances of the data are of equal importance. The globally correlated data are reduced using the magnitude or rank-over to choose the optimal low-dimensional pairwise distance, thereby eliminating irrelevance and redundancy [3]. With the LDR technique, only the local distances are assumed to be reliable in high-dimensional space. More emphasis is therefore put on correctly modeling the locally correlated data to eliminate the noise [4]. A great deal of research has been conducted using realistic nonimage and image datasets, with the results showing that a significant percentage of redundancy and noise is eliminated by both the techniques [5–7].
Recently, a number of hybrid global-local methods have been proposed. The authors of [5], for instance, have developed an algorithm of hybrid sampling-based clustering ensemble with global constitution, which encodes the local and global cluster structure of input partitions in a single representation. Here deciding a final consensus candidate involves significant computational cost. In [6], a DR algorithm was proposed which uses pairwise similarity measurement to effectively capture the local structure of data manifolds. While offering considerable advantages, the hyperparameters of the selected model remain as an issue. In [7], a global-local structure preservation framework was introduced for feature selection based on three algorithms: local linear embedding (LLE), local tangent space alignment (LTSA), and locality preserving projection (LPP). In this framework, the copious amount of noise generated during the process may degrade the reliability of the data. A well-designed DR model still needs to be developed that can effectively reduce the data redundancy and noise while ensuring robustness.
In this paper, thus, a global-local DR model called sparse locality for PCA (SLPCA) is proposed, which introduces robustness into both GDR and LDR by employing a regularizer and constraint, separately. The SLPCA model aims to reduce unnecessary information while preserving the data correlation in both the global and local structure. In order to effectively consider the trade-off between the complexity and efficiency, a robust L2,p-norm constraint-based PCA (R2P-PCA) is introduced for GDR, while sparse representation (SR) and LPP are combined for LDR. The R2P-PCA algorithm implements PCA with a robust distance metric L2,p-norm by maximizing the sum of the variations. In this way, R2P-PCA can increase the robustness against the noise. It is also for preserving the global structure of the samples. The SR-LPP algorithm seeks a set of projection matrices by capitalizing the merits of SR and LPP, which are merged into one analytical process. SR enables adaptive construction of the graph because it is parameter-free and robust against the noise [8]. This makes SR-LPP capable of simultaneous and adaptive learning of the projection matrix and sparse similarity matrix. Through the joint learning of the two algorithms, the learned functions can be optimized using an efficient iterative algorithm, alternating direction method of multipliers (ADMM) [9]. It monotonically decreases the value of the augmented Lagrangian function as the iteration continues. Computer simulation reveals that the proposed SLPCA scheme consistently outperforms the existing schemes in terms of clustering accuracy and data reconstruction error. The main contributions of the paper are summarized below:(i)A means of delivering robustness in DR is developed that simultaneously considers both the local and global structure of sample data.(ii)A scheme effectively considering the trade-off between the complexity and efficiency of DR is proposed, where a robust L2,p-norm constraint-based PCA (R2P-PCA) is introduced for global DR, while sparse representation (SR) and LPP are combined for local DR.(iii)Typically, the graph-based DR methods calculate a projection matrix on the basis of a learned graph. A novel approach is developed that can simultaneously integrate both the projection matrix and nonparametric graph construction.
The remainder of the paper is organized as follows. In Section 2, the work related to the preservation of global and local data structure is surveyed and summarized. The proposed sparse locality for PCA (SLPCA) model is introduced in Section 3. This section also presents the key algorithms of SLPCA, R2P-PCA, and SR-LPP and explains how they achieve convergence. In Section 4, the performance of SLPCA is evaluated in terms of clustering accuracy and data reconstruction error with two kinds of datasets. The paper is concluded in Section 5, with future research directions.
2. Related Work
2.1. Notation and Definition
Given the vector , the Lp-norm of is defined as =( )(1/p). When , the L0-norm of is defined as . When , the L1-norm of vector is . Given the matrix Mij ∈ Rn×m, the i-th row and j-th column are denoted by mi and mj, respectively. L0-norm ||M||0 denotes the number of nonzero elements in the matrix M. The L1-norm of M is defined as
The Frobenius norm of M is
The L2,1-norm of M was first introduced in [10] as a rotational invariant for the rows. It is now widely employed to encourage row sparsity and is defined as
2.2. DR Based on Global Structure
There exist three main approaches for the DR based on the global structure in the literature.
2.2.1. Principal Component Analysis
PCA is the most commonly used algorithm in this category because of the simplicity and efficiency. It aims to reduce the dimension of the original data by projecting them across the direction of maximum variance. The Graph-Laplacian PCA algorithm was developed by Jiang et al. [11], which learns a low-dimensional representation of vector data by incorporating the graph structure encoded in the graph data. In [12], rotational invariance L1-norm PCA minimizes the reconstruction error by imposing an L2-norm on the spatial distance, whereas an L1-norm is applied to different data points. The authors of [13] have proposed the convex sparse PCA approach that reduces redundant information by building a compact and informative subspace. These kinds of algorithms are useful in the fields including image processing, speech enhancement, and reduction of data transmissions.
2.2.2. Distance Preserving Method
The typical distance preserving algorithms are based on maximum variance unfolding (MVU) and isometric mapping (Isomap). Although both the approaches result in a similar computational structure, Isomap merely preserves the global geodesic distance, whereas MVU preserves the local distance but maximizes the global variance. In [14], a hybrid incremental landmark MVU model was developed that was combined with a dual-tree complex wavelet transform. Saxena et al. [15] introduced a nonlinear MVU-PCA algorithm, which aims to enhance the core components of the framework for higher robustness. The multilevel MVU technique proposed in [16] aims to reduce and distribute the computational load across parallel multilevel machines. The algorithms mentioned above share a weakness of erroneous graph connection.
2.2.3. Autoencoder
Autoencoder (AE) [17] focuses on learning the underlying manifold in the encoding and decoding procedure. The multilayer AE [18] usually has a higher number of connections, but it may cause slow convergence via backpropagation even to local minima. The spectral AE [19] aims to detect global and community anomaly by mapping the attributed network to two types of low-dimensional representations. The AE-based approaches are rather limited in their capacity in learning the local structure, which contains important information required for understanding the data.
2.3. DR Based on Local Structure
The DR schemes based on local data structure usually adopt the graph embedding technique such as LPP, LLE [20], Laplacian eigenmaps [21], LTSA [7], and neighborhood preserving embedding (NPE) [22]. They approximate the embedded manifold through a mapping function with the proximity preserving property. The LPP algorithm uses the k-NN-based graph Laplacian regularizer to preserve the local structure of the samples. Yu et al. [23] improved the LPP scheme where the L1-norm is used to provide the robustness while the similarity between the pairs of vertices is effectively preserved simultaneously. In [24], a supervised global-locality preserving projection approach is proposed that preserves not only the locality characteristics of LPP but also the global discriminative structure associated with maximum margin criterion. The aforementioned algorithms can be hampered by the presence of noise. The low-rank learning method can help to reduce the disturbance caused by noise in the data.
In [25], a neighborhood preserving projection method was developed that integrates low-rank learning and robust learning. It enhances the robustness of the NPP method and diminishes the disturbance caused by noise in the data. Similarly, the low-rank preserving projection method introduced in [26] applies an L21 norm to the noise matrix as a sparse constraint and a nuclear norm to the weight matrix as a low-rank constraint. It maintains the global structure of the data during DR, and the learned low-rank weight matrix reduces the noise-related disturbance. The scheme of [27] preserves the structure of nonlinear manifold data, and it is robust because the learned data are unaffected by noises or outliers. In [28], a structurally incoherent low-rank nonnegative matrix factorization method was proposed that jointly considers the structural incoherence and low-rank property of data. Since the scheme employs low-rank learning method, it allows to capture the global structure of the data and robustness to the noise. The proposed scheme is presented next.
3. The Proposed Scheme
3.1. R2P-PCA for Preserving Global Structure
R2P-PCA is for the preservation of the global structure of the data. PCA finds a few orthogonal directions in the data space that preserve the most information in the data while minimizing the reconstruction error. Assuming an input data matrix, X=(x1, …, xn)∈Rm×n, that contains m data column vectors in n dimensional space, PCA finds the optimal low-dimensional subspace of U and by solving the following optimization problem:where each column of U (= (u1, …, uk) ∈ Rm×k) is a principal direction. (= ∈ Rk×n) represents the projected data points in the new reduced subspace.
Recently, various PCA-based methods have employed different criterion functions such as L1-norm to improve the robustness to noise [29]. Although L1-norm is robust to outliers, most existing L1-norm-based PCA methods do not effectively minimize the reconstruction errors which is one of the main goals of PCA. They are also not invariant under rotation, which is an important property for learning algorithm. In the proposed scheme, we therefore focus on formulating a robust learning metric formulation, the L2,p-norm [30–32]. Note that most existing PCA methods based on L2,1-norm can be viewed as special cases of the proposed R2P-PCA approach. It is well known that ||·||2,1 is convex with respect to the matrix variables, and thus it can be extended to a generalized robust learning metric formulation for matrix M, namely, the L2,p-norm (0 < p < 2), which can be defined as follows:
On the basis of equations (4) and (5), the objective function of R2P-PCA can now be obtained:
3.2. SR-LPP for Preserving Local Structure
3.2.1. Original Locality Preserving Projection
LPP can be viewed as a linear approximation of nonlinear Laplacian eigenmaps. The first step in the original LPP is to construct an adjacency graph, which significantly affects the performance of the algorithm. Then, the projection matrix is calculated using the learned graph. The objective function of LPP can be formally stated as follows:
Let denote the transformation vector. Through simple algebraic reformulation, the objective function above can be rewritten as follows:where pTxi, , and are numerical values, Dii = ∑j, and L = . Here, -L denotes the Laplacian matrix, which is used to minimize PTXLXTP. The LPP objective function can therefore be changed to
For the heat kernel method, the weight assignment can be obtained using
In [33], it was established that adjacent graph structure and graph weight are highly interrelated and should not be separated. This makes it preferable to develop one ideal model that can perform the two tasks simultaneously. The traditional weight assignment methods require parameter selection to construct the adjacency graph by means of the ɛ-ball or k-NN method. This comes with a significant computational cost if they need to identify the neighborhoods for a whole dataset. This also makes them sensitive to data noise [34]. Hence, instead of using them, here we attempt to automatically derive the similarity matrix and ensure that it preserves the discriminative information by using sparse representation (SR).
3.2.2. Graph Construction Based on SR
The main idea of SR is that a body of sampled data can be sparsely represented when given an appropriate basis. If a given test sample, y, belongs to the i-th class, SR assumes that y can be represented as a linear combination of the training samples in the i-th class, Xi = {xi1, xi2, ..., xiNi}. In other words, y can be represented as , where denotes the coefficient of y over Xi. Ideally, the coefficients of other classes are zero, such that = 0 and j ≠ i. Thus, y can be represented as a linear combination of all the training samples by stacking Xi (i = 1,2, ..., C) within the whole training set, X. Then, the coefficient matrix, , can be obtained. So,
The above model can be expressed as follows:
Note that the L0-norm optimization problem is an NP-hard problem. A recent study has demonstrated that L0-norm is equivalent to the L1-norm optimization problem if the solution is sparse enough [35], in which case it can be solved using
SR avoids parameter selection and makes the intrinsic graph construction more robust to data noise. Having established the ground of theoretical inspiration related to LPP and SR, the basic idea of the proposed SR-LPP method for local structure preservation is presented next.
3.2.3. SR-LPP
SR-LPP learns the similarity matrix, , and projection matrix, P, simultaneously and adaptively so that the intrinsic properties of the structure can be preserved. For this, the LPP objective function needs to be changed to
In addition, in order to ensure that each sample is represented by unique base, an L1-norm constraint is added to . According to equation (9) in [36], the objective function for SR-LPP can now be formulated as follows:where λ1 and λ2 (≥0) are parameters balancing the contribution from the two parts.
3.3. The Objective Function
For joint learning with R2P-PCA preserving global structure and SR-LPP preserving local structure, equations (6) and (15) are combined in the following model:where α (≥0) is a parameter balancing the contribution from the three parts. It should be pointed out that the update of P in equation (16) can be derived by solving the generalized eigenvector problem, which corresponds to the eigenvectors of the k smallest eigenvalues. Let = PTX. As in PCA takes exactly the same role as pTxi in LPP, the objective function can be simplified towhere = . ADMM can be expanded to solve the subproblem. Letting , the standard Lagrange function in the equation above can be augmented as follows:
Hence, the iterative computation consists of , U,, and -minimization step and an update of the Lagrangian parameter, ρ. The detailed iterative computational steps are formulated below.
3.4. Optimization Analysis
3.4.1. Fixing and W to Update U and
According to Theorem 3.1 in [11], let ; U can be updated by solving the following optimization problem:
Taking the partial derivative of U and setting it to 0, the following can be obtained:
Then, U is updated. As , the optimization problem with respect to in equation (18) can be simplified into the following equation:With similar operation on U,
The optimal can be obtained by calculating the eigenvectors, = . This is the first k smallest eigenvalue of the matrix According to Proposition 3.2 in [11], the solution of can be expressed by the eigenvectors ofwhere σm and σl are the largest eigenvalue of matrix MTM and L, respectively; θ is the parameter substituting α; and e (= (1…1)T) is an eigenvector of . Although is applied in place of the Laplacian matrix, the eigenvectors and eigenvalues do not change. This guarantees that e is not included in the lowest k eigenvectors.
3.4.2. Fixing U, V, and W to update
Let . can now be updated by solving the following optimization problem:
Taking the partial derivative of and setting it to 0, according to the proximal operator of L2,p norm proposed in [30], change to , where is the i-th column of and 0 < ɛ0 ≤ 1. Then, can be obtained usingwhere N is the weight matrix corresponding to ; here, . The L2,p-norm minimization problem can then be solved by updating N and iteratively. The procedure for this is presented in Algorithm 1.
|
3.4.3. Fixing U, V, and to Update W
can be updated by solving the following optimization problem:
Taking the partial derivative of and setting it to 0, apply the nonnegative projection proposed in [37]. Then, obtain using
3.4.4. Updating the Parameters
Drawing upon [9], the parameters need to be updated as follows:
3.5. Convergence Property
The proposed sparse locality-regularized PCA model (SLPCA) consists of two main parts: solving the subproblems and updating the parameters. For the update of the first subproblem, the optimal solution of each column of is obtained using equation (24). Therefore, the value of the objective function decreases after solving in each iteration. For the other two subproblems, the minimization of the augmented Lagrangian occurs with respect to the variables, U, , and . When updating U, because the Lagrange function of equation (18) is also convex and differentiable to U, the optimal projection matrix can be obtained by using a closed-form solution of equation (20). This means that the value of the objective function decreases after solving U in each iteration. The same procedure is applied for updating and . The proposed SLPCA algorithm is therefore convergent.
The procedure for the proposed SLPCA is shown in Algorithm 2, which is applied when the subproblems are solved. It searches for an optimal value in each iteration and finally stops when the value of the objective function becomes stable.
|
3.6. Computational Complexity
With the proposed SLPCA algorithm, the major computational burden lies in Steps 2 and 4 of Algorithm 2. Step 2 is for the eigenvalue decomposition on an m × m matrix, of which computational complexity is O(m3). Step 4 is the inverse operation of an n × n matrix of the complexity of O(n3). Therefore, the computational complexity of Algorithm 2 is O(t(m3 + n3)), where t is the number of iterations.
4. Performance Evaluation
4.1. Experiment Setup
This section reports an experiment conducted to verify the effectiveness of the proposed SLPCA scheme. A comparison is also made regarding the performance in finding low-dimensional representation and classification in relation to the two baseline algorithms, PCA and LPP, and three state-of-the-art hybrid global-local DR methods, i.e., GRSDA [38], L1/2-GLPCA [39], and p-GLPCA [40]. In the experiment, the value of the parameter θ is chosen from the set [0.1, 0.5, 0.9], and is chosen from the set [0.5, 1]. According to the result of experiment, the proposed method is superior to the others when the parameter α is larger than 0.6. The value of λ1, λ2, and ρ are empirically determined, while it was found that their values do not significantly influence the performance.
In order to evaluate the performance of the schemes on various characteristics including classification and data reconstruction accuracy, the experiment is performed using two different types of datasets: nonimage-based dataset and image-based dataset. For the nonimage data, three UCI datasets are employed: Iris, Seeds, and Soybean that are publicly available at https://archive.ics.uci.edu/ml/datasets.php. For the image data, three commonly adopted datasets are used, the Extended Yale-B, ORL face, and CMU PIE dataset. The K-means test is applied to assess the effectiveness of the proposed scheme. The experiment is also extended to test its robustness and generalizability for different values of . Table 1 summarizes the main features of the DR models compared.
4.2. Experiment on Nonimage Datasets
This section details the results of the experiment conducted to assess the effectiveness of the schemes in terms of subspace representation and clustering accuracy. The Iris dataset is one of the well-known pattern recognition databases published in the literature. It contains three classes of 50 instances each, where each class refers to a type of iris plant. Here, one class is linearly separable from the other two classes, which are not linearly separable from each other. The Seeds dataset covers three different varieties of wheat—Kama, Rosa, and Canadian—each containing 70 elements, which were randomly selected for the experiment. The Soybean dataset has 19 classes, only the first 15 of which have been used in prior work. It contains 35 categorical attributes, some nominal and some ordered.
By using equation (5), L2,p norm can be applied in the SLPCA algorithm to improve the robustness. The previous research has established that L2,p-norm (0 < p < 2) performs better than other constraints [32]. To apply it, some parameters need to be set in advance, including the rescaling coefficient. Different values are tried while other parameters are unchanged to find the optimal value. The other parameters are then decided in the similar way. Figure 1 shows the distribution of 2D data after DR with the Iris dataset. In this experiment, α = 0.6, λ1 = 0.5, λ2 = 0.5, ρ = 1.0, , and θ = 0.1. For the original LPP, the neighborhood size, k, also needs to be determined, and k = ni = 1 where ni is the number of samples in the i-th class. It can be seen from the figure that the proposed SLPCA scheme represents the Iris data pattern more compactly than the other algorithms.

(a)

(b)

(c)

(d)

(e)

(f)
The clustering accuracies for the training set and test set of the schemes are compared in Table 2. The best results are highlighted in bold. Note that outside of a few exceptional cases, the proposed SLPCA scheme consistently displays the highest accuracy for both the training and test data, across the different datasets.
4.3. Experiments Using Image Datasets
In this section, the performances of the different schemes are compared with regard to image classification and reconstruction. The experiments were conducted on three face datasets: CMU PIE, Extended Yale-B, and ORL. For each image, salt and pepper noise were added, with a noise density of 0.01.
4.3.1. CMU PIE Dataset
The CMU PIE dataset is publicly available at http://www.cs.cmu.edu/afs/cs/project/PIE/MultiPie/Multi-Pie/Home.html. It has 41,368 images of 68 individuals, with 4 expressions, 13 poses, and 43 illumination conditions for each person, and all the images have different illuminations and expressions. In our experiment, the Pose C09 subset was chosen, which contains 1632 images of 68 people, with 24 different images of each person. In the experiment, all the images are in gray scale and normalized to a resolution of 64 × 64 pixels for the sake of efficient computation. 70% of the CMU PIE images are used as the training dataset. The rest forms the test dataset.
4.3.2. ORL Face Dataset
The ORL face dataset consists of 400 images of 40 individuals, with 10 different images per person. It is publicly available at http://www.face-rec.org/databases/. The images of each person were taken at different times with different facial expressions and details and under varying lighting conditions. In the experiment, all the images are in gray scale and are manually cropped to 64 × 64 pixels. For this dataset, 60% of the images per subject were randomly chosen to form a training set. The remaining images are then used as the test set. For each algorithm, the number of iterations is set to 20 and the best recognition rate is selected along with the corresponding residual error.
4.3.3. Yale-B Face Dataset
The Yale-B face dataset is also popular in face recognition research. It is publicly available at http://vision.ucsd.edu/∼leekc/ExtYaleDatabase/Yale%20Face%20Database.htm. This dataset consists of 2432 face images of 38 subjects under various lighting conditions. 64 gray scale images were selected for each subject, and each image was resized to 32 × 32 pixels. In this experiment, 10, 20, and 30 images per subject were chosen to form the training set, with the remaining images being formed as the test dataset. The same procedure was applied as with the ORL dataset.
To illustrate the data representation aspect of SLPCA, ORL face dataset is applied here. In this experiment, α = 1.0, λ1 = 1.0, λ2 = 1.0, ρ = 0.7, and θ = 0.9. Figure 2 shows the original images (in 1st column) and reconstructed images of L1/2-GLPCA, p-GLPCA, and SLPCA. The 2nd column shows the reconstruction images of L1/2-GLPCA while the 3rd and 4th column are the results of p-GLPCA at . The last three columns contain the reconstructed images of SLPCA with . Only the images of three ORL persons (out of 40 persons) are shown due to space limitation. Observe from the figure that the images of a person with the proposed SLPCA scheme tend to be more similar than those with the other methods. This indicates that the class structures are generally more apparent with the SLPCA representation, and this motivates us to use it for data clustering and classification.

Table 3 compares the clustering accuracy for the GLPCA and SLPCA schemes for different p values. Here, L1/2-GLPCA and p-GLPCA used an L2,1-norm and a p-norm for the PCA aspects, respectively. SLPCA used an L2,p-norm. It can be seen from the table that SLPCA (with ) achieved the best performance out of all the schemes. In this experiment, α = 1.0, λ1 = 1.0, λ2 = 1.0, ρ = 0.7, , and θ = 0.9. Figure 3 compares the schemes with regard to the reconstruction error and residual error () when the subspace dimension was changed for the 30% CMU PIE training dataset. It can be seen that SLPCA with a proper value produced the smallest reconstruction and residual error rate.

(a)

(b)
Figure 4 compares the error rates for different θ values. This reveals that SLPCA has a lower reconstruction error for any value of and the lowest residual error when , except for p-GLPCA where . Note that in the case of SLPCA, the reconstruction error decreases as θ becomes larger, while the residual error increases slowly. Figure 5 shows a comparison of the schemes across the three different image datasets. It can be seen that 20% CMU PIE and ORL datasets have the lowest and largest reconstruction error rate, respectively. It seems that there is no relationship between the two types of error rates for the given datasets. Here, α = 1.0, λ1 = 1.0, λ2 = 1.0, ρ = 0.5,, and θ = 0.5.

(a)

(b)

(a)

(b)
From the experiment results given above, SLPCA turns out to be the most robust for handling DR for various types of data because it retains both the local and global structure of the data.
5. Conclusion
In this paper, a robust global-local scheme called SLPCA has been proposed for dimensionality reduction. It increases robustness against noise and more effectively preserves the global and local structure of the samples. In seeking a trade-off between complexity and efficiency, robust L2,p-norm-based PCA (R2P-PCA) was introduced for the GDR elements, while joint sparse representation and LPP (SR-LPP) was used for the LDR. In addition, the SR-LPP algorithm is parameter-free so as to avoid the difficulty of determining the neighborhood size. This allows the graph construction and weight assignment to be finished in one step. SR-LPP can also learn a projection matrix and a sparse similarity matrix simultaneously and adaptively. Experimental results show that the proposed scheme is capable of more accurate classification and better representation than other typical DR schemes.
In future work, the proposed scheme will be extended to take nonlinear settings into account. As a linear model cannot capture nonlinear distortion, a nonlinear model is required to obtain effective and reliable results. The identification of nonlinear models requires more data and involves more comprehensive analysis than their linear counterparts. At present, the performance of the representation has been studied only in linear settings. How to make the linear models work in nonlinear settings is another issue that requires further investigation. Analytical models deciding the values of the parameters employed in a DR scheme also need to be developed.
Data Availability
All data included in this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Acknowledgments
This work was partly supported by the Institute for Information and Communications Technology Promotion (IITP) grant funded by the Korea government (MSIT) (No. 2016-0-00133, Research on Edge Computing via Collective Intelligence of Hyperconnection IoT Nodes), Korea, under the National Program for Excellence in SW supervised by the IITP (Institute for Information and Communications Technology Promotion) (No. 2015-0-00914) and Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (2017R1A2B2009095, Research on SDN-based WSN Supporting Real-time Stream Data Processing and Multiconnectivity, and 2019R1I1A1A01058780, Efficient Management of SDN-based Wireless Sensor Network Using Machine Learning Technique), the second Brain Korea 21 PLUS project.