1 Introduction

Modern data can be overwhelming to interpret due to their size and dimensionality. As a result, the demand for understanding these data is growing rapidly. For analyzing high-dimensional data (Yuan et al. 2013) such as human facial expression images (Zhang et al. 2003), visualization techniques can provide instinctive knowledge of data. A normally used method to achieve an accurate visualization of high-dimensional data is learning a low-dimensional embedding of the high-dimensional data (Fujiwara et al. 2011). The low-dimensional representation of data should reveal corresponding relationships in higher dimensions. Specifically, data in close proximity represent similarity and data separated by long distances represent dissimilarity.

Conventional visualization methods derive from dealing with the problem of dimensionality reduction. Different methods are proposed in dimensionality reduction techniques, such as principal components analysis (PCA) (Jolliffe 1986) and nonnegative matrix factorization (NMF) (Lee and Seung 2001). Matrix transformations are taken to obtain the principal components in a smaller matrix fulfilling dimensionality reduction. In general, matrix operations are easy to be realized and can provide results quickly. However, these dimensionality reduction approaches are not capable of preserving the dimensionality information except the principal components. One can expect a general dimensionality reduction result but at the expense of meticulous embedding.

Multi-dimensional scaling (MDS) techniques (Cox and Cox 2001) are commonly used as a means of visualizing data while preserving dimensionality information in the form of distances. A set of MDS techniques can be found in the literature, such as Isomap (Tenenbaum et al. 2000), locally linear embedding (LLE) (Donoho and Grimes 2003), Sammon mapping (Sammon 1969), and LAMP (Joia et al. 2011). In general, an MDS algorithm is designed to place data iteratively in low-dimensional space such that the distances between data are preserved as well as possible. The majority of these techniques attempt to simulate the short pairwise distances between data, which are considered to be dependable in high-dimensional space. For example, LLE considers preserving only the local, small distances at the expense of not including remaining distances. Furthermore, there is much uncertainty as to what defines the “local” range. Even though the famous Sammon mapping method optimizes all the mutual distances (i.e., not just local small, local distances), it suffers from many overlaps between categories and is not guaranteed to converge. LAMP is one of the MDS techniques based on landmarks. Different landmarks give different MDS results. In summary, traditional MDS methods still have the following shortcomings:

  1. 1.

    The optimization on distance preservation uses only the first-order gradient method, which is easily trapped in local minima.

  2. 2.

    Narrow margins among clusters create overlapping in the mapped results.

  3. 3.

    Most existing methods only consider distances between data in embedding, whereas additional factors could provide desirable information.

  4. 4.

    Connected relations are not shown in the embedded results.

In this paper, we revisited MDS techniques and asked the question: Can we use an unsupervised learning approach to conduct MDS purpose for both data visualization and clustering while discovering data relations? In order to improve the stated shortcomings of traditional MDS methods, optimization-based MDS with unsupervised clustering may provide both a promising and effective solution. Optimization methods using second-order gradient descent have the ability to produce more accurate results than first-order methods due to their stronger ability to escape from local minima in our case. However, second-order gradients often bring heavy computation load. In unsupervised learning, clustering is one of the most advanced techniques that can provide data category information. Regarding what the users want to get from the MDS results, cluster automatic formation is thought to be one of the highest demanded expectations. Other expectations include larger cluster margins and individual data relationships. Therefore, in order to utilize the second-order gradient approach to fulfill MDS purpose with cluster automatic formation as well as possible, this paper proposes a new density-concentrated multi-dimensional scaling algorithm for relation visualization, called DCMDS-RV. The key idea behind the proposed DCMDS-RV algorithm is to use density-based clustering and incorporate it with Levenberg–Marquardt (LM) optimization-based MDS. This algorithm presents an alternative MDS approach using LM optimization and density concentration, yielding improved MDS performance. The main contributions of this paper are summarized as follows:

  1. 1.

    We propose a new unsupervised algorithm for general MDS purpose based on LM optimization method. As LM method can automatically switch between first-order gradient and second-order gradient optimization methods, the MDS technique using LM optimization shows great improvement in mapping data because of its ability to escape from local minima.

  2. 2.

    To obtain a better visualization on data category information, density-based clustering is integrated in the MDS process as an auxiliary methodology. In the mapping results, data move based on their mutual distances as well as their density relationships. Because of this, better cluster gathering and larger cluster margins are achieved during mapping without knowing any category knowledge.

  3. 3.

    Our proposed algorithm is evaluated and compared with other MDS approaches, including Sammon mapping, Isomap, LLE, and LAMP. When evaluated on experiments involving mapping several real-life data on a 2D plane, DCMDS-RV algorithm outperforms traditional MDS approaches. Moreover, it has the ability to provide mapping results in any desired dimensional space.

  4. 4.

    In order to reveal the relations between data, nearest neighbor map (NNM) is generated along the algorithm and no extra step is needed. The relations between data are shown in the embedded results as connected lines, providing a more vivid and intuitive graph view of data.

The rest of this paper is organized as follows: Section 2 briefly reviews the related concepts. Section 3 details our proposed DCMDS-RV algorithm for data relation visualization. Section 4 demonstrates the experimental results on several high-dimensional datasets to evaluate the effectiveness of the proposed algorithm. Finally, Sect. 5 draws the conclusions.

2 Related works

2.1 Overview of MDS approaches

MDS is a technique used for embedding high-dimensional data into a low-dimensional space where the distances in the low dimension well represent the distances in the original, high-dimensional space. As data with short distances are similar to each other, MDS can also be applied to analyze the similarity, dissimilarity, and relationships between data. The general cost function of MDS onto a 2D plane is defined in (1), where n is the data number. The general goal of MDS approaches is to minimize the cost function.

$${\text{Err}} = \mathop \sum \limits_{i = 1}^{n - 1} \mathop \sum \limits_{j = i + 1}^{n} \left( {\sqrt {\left( {x_{i} - x_{j} } \right)^{2} + \left( {y_{i} - y_{j} } \right)^{2} } - d_{ij} } \right)^{2}$$
(1)

where xi and yi are the x and y coordinates of mapped points i on 2D plane, respectively. dij is the distance calculated in the original high-dimensional space.

In general, there exist two types of MDS algorithms: metric and non-metric.

  1. 1.

    In metric MDS approaches, the actual values of the dissimilarities are used. The distances between data points are then set to be as close as possible to the similarity or dissimilarity data. Sammon mapping and LAMP techniques are two of the standard metric MDS approaches.

  2. 2.

    In non-metric MDS approaches, the order of the distances is preserved, and instead a monotonic relationship between the similarities/dissimilarities and the distances in the embedded space is found. In most non-metric MDS techniques, such as Isomap and LLE, criterion choices become undefined when two data points are at the same location (“colocated”), which means there are same distances between some data. When “colocated” happens, it will suspend non-metric MDS processing without an MDS solution.

2.2 Density-based clustering

Given the assumption that cluster centers are surrounded by neighbors with lower local densities and are at a relatively large distance from any data with a higher local density, density-based clustering methods (Bo and Wilamowski 2017; Rodriguez and Laio 2014) are capable of fulfilling unsupervised clustering task with satisfying results. There are three steps according to density-based clustering approaches (Rodriguez and Laio 2014).

Step 1: Calculate the data local density ρi. Regarding density computation, a Gaussian kernel-based density contribution is given as (2).

$$\rho_{i} = \mathop \sum \limits_{i \ne j} { \exp }\left( { - \left( {\frac{{d_{ij} }}{{d_{\text{c}} }}} \right)} \right)^{2}$$
(2)

here ρi is the local density of data i. dij is the distance between data i and j. dc is the cutoff distance.

Step 2: Calculate the minimum distance δi between data i and any other data with a higher local density. The calculation of δi is given in (3).

$$\delta_{i} = \left\{ {\begin{array}{*{20}l} {\hbox{min} \left( {d_{ij} } \right),} \hfill & {\rho_{j} > \rho_{i} } \hfill \\ {\hbox{max} \left( {d_{ij} } \right), } \hfill & {\rho_{j} < \rho_{i} } \hfill \\ \end{array} } \right.$$
(3)

here \(j = 1,2, \ldots ,{\text{np}}\;{\text{and}}\;j \ne i\).

Step 3: Generate the decision graph. Aim to choose the data with large local density ρi as well as large minimum distance δi as the cluster centers. The decision graph generated with x-coordinate is ρi and y-coordinate is δi. After the selection of cluster centers in the decision graph, data will be assigned into different clusters based on the minimum distances δs.

3 Proposed density-concentrated multi-dimensional scaling for relation visualization (DCMDS-RV)

The proposed algorithm focuses on projecting high-dimensional data into a low-dimensional (2D/3D) space with data relations revealed. Overall description of the proposed DCMDS-RV is illustrated in Fig. 2. The benefits of the proposed DCMDS-RV algorithm include applying first-order and second-order gradient descent methods to optimize data locations and using density relationships between data to concentrate clusters and enlarge margins between them. It simultaneously learns MDS embedding and clustering. Moreover, data relations in density field are found during the process, which can provide us a better understanding of the original data and their relations.

3.1 MDS process in the proposed DCMDS-RV

The cost function of MDS methods in (1) leads to an alternating nonlinear least squares optimization process, where we alternate between re-computing different data, and each step and iteration is guaranteed to lower the value of the cost function. In most cases, the optimization process is fulfilled using first-order instead of second-order gradient approaches, allowing the process to be trapped in local minima. Levenberg–Marquardt (LM) method, which is developed to solve nonlinear least squares problems iteratively, finds the best solutions by switching between first-order and second-order gradient approaches via a damping parameter. Unlike the second-order gradient methods, which are of heavy computation, LM method approximates the second-order gradient with the first-order gradient. In this paper, LM method is adopted as the MDS technique to determine data positions. In the proposed DCMDS-RV algorithm, MDS embedding process has two phases: (1) data position initialization using matrix eigen-decomposition and (2) data position optimization using LM method.

3.1.1 Getting initial data positions for LM method via matrix eigen-decomposition

Instead of randomly generating initial positions, the matrix eigen-decomposition technique is used to provide initial positions very fast on a 2D plane or 3D space. It can accelerate the nonlinear least squares optimization process.

Suppose DMat is the distance matrix that contains all the between-data distances and DMat(ij) returns the distance between data i and j. Then, the initial data positions on a 2D plane are given as (10) via matrix eigen-decomposition.

$${\text{SD}} = {\text{DMat}} .*{\text{DMat}}$$
(4)
$${\text{ti}} = {\text{sum}}\left( {{\text{SD}}\left( {i,:} \right)} \right)$$
(5)
$${\text{tj}} = {\text{sum}}\left( {{\text{SD}}\left( {:,j} \right)} \right)$$
(6)
$${\text{ta}} = {\text{sum}}\left( {{\text{sum}}\left( {\text{SD}} \right)} \right)$$
(7)
$$M\left( {i,j} \right) = \frac{1}{2}\left( {{\text{ti}} + {\text{tj}} - {\text{ta}} - {\text{SD}}\left( {i,j} \right)} \right)$$
(8)
$$\left[ {V,D} \right] = {\text{eigs}}\left( M \right)$$
(9)

More details on eigen operations can be found in Abdi (2007). Diagonal matrix D contains the eigenvalues on the main diagonal, and the columns of matrix V are the corresponding eigenvectors. Initial data positions in a 3D space are shown as (11). Note that the first eigenvalue D(1, 1) is zero, which is neglected here.

$$2 {\text{D}}\;{\text{mapping:}}\;P^{\left( 0 \right)} = \sqrt {D\left( {2:3,2:3} \right)} *V\left( {:,2:3} \right)^{\prime }$$
(10)
$$3 {\text{D}}\;{\text{mapping:}}\;P^{\left( 0 \right)} = \sqrt {D\left( {2:4,2:4} \right)} *V\left( {:,2:4} \right)^{\prime }$$
(11)

Next, with the initial positions, embedded data are optimized iteratively using LM method.

3.1.2 LM method

LM method (Marquardt 1963) is for minimizing the cost function (1) in DCMDS-RV algorithm. In order to return better gradient-based optimization results, the second-order derivatives of the total error function are considered. However, the calculation of Hessian matrix H, which contains the second-order derivatives of cost function, is often complicated. In order to simplify the computing process (Wilamowski and Yu 2010), the Jacobian matrix J is introduced to approximate Hessian matrix H. J is the matrix of all first-order derivatives of the cost function with respect to data’s coordinates. For the cost function (1) of 2D mapping, the m’th row of Jacobian matrix is \(J_{m :} = \left[ {\frac{{\partial {\text{Err}}}}{{\partial x_{m} }} \frac{{\partial {\text{Err}}}}{{\partial y_{m} }}} \right]\).

$$H \approx J^{\text{T}} J$$
(12)

In order to make sure that the approximated Hessian matrix JT J is invertible, LM algorithm introduces another approximation to Hessian matrix:

$$H \approx J^{\text{T}} J + \mu I$$
(13)

where μ is combination coefficient with positive value; I is the identity matrix with the size of 2 × 2 for 2D MDS. From Eq. (13), one may notice that the elements on the main diagonal of the approximated Hessian matrix will be larger than zero. Therefore, with this approximation (13), it can be sure that matrix H is always invertible.

Now, the update rule of LM algorithm for 2D MDS case can be presented as (14)–(16):

$$\Delta = \left( {J^{\text{T}} J + \mu I } \right)^{ - 1} *J*{\text{Err}}$$
(14)
$$x = x + \Delta \left( 1 \right)$$
(15)
$$y = y + \Delta \left( 2 \right)$$
(16)

As the combination of the steepest descent and second-order gradient algorithms, LM algorithm switches between the two algorithms during the least squares minimization process. When the combination coefficient μ is very small (nearly zero), (14) approaches the second-order algorithm; when the combination coefficient μ is very large, (14) approaches the steepest descent method.

3.2 Density concentration and relations generation process in the proposed DCMDS-RV

Density-based clustering methods assume that cluster centers are surrounded by neighbors with lower local densities. In other words, data with small local densities should move close to data with large local densities when embedding. The general new idea behind this is the density concentration process, where each datum will be mapped closer to its nearest neighbor data in the density field.

Step 1: Calculate the data local density ρi using (2).

Step 2: Generate nearest neighbor map (NNM) in the meantime of calculating the minimum distance δ (3). Nearest neighbor information will be stored in NNM, where data connect to its nearest neighbor of larger local density, with their distance equal to δ. For instance, if the minimum distance δ for data point m is δm from data point m to data point n, then data point n is the nearest neighbor of data point m in the density field. Notice that data’s nearest neighbor in NNM always has a larger local density, with their distance as δ. The datum with the largest local density connects to itself in the NNM. This also generates the relations between data in the density field, which are shown connected in the resulting embedding. An NNM example of a two-dimensional dataset is shown in Fig. 1.

Fig. 1
figure 1

An illustration example of NNM for two-dimensional dataset-FLAME (Fu and Medico 2007). The colors of data stand for different density values. In NNM, data’s nearest neighbor always has a larger local density, with their distance as δ

With the involvement of NNM, it is possible to present data structures as graphs. In a variety of research and application areas, data graph modeling and analysis (Kennedy et al. 2017) are an important task to be accomplished, which makes graph visualization (Cheng et al. 2018) necessary when developing visualization techniques.

$${\text{Moving}}\;{\text{rate:}}\;\lambda = 1 - \frac{\delta \left( i \right)}{{{ \hbox{max} }\left( \delta \right)}}*\frac{\rho \left( i \right)}{{{ \hbox{max} }\left( \rho \right)}}$$
(17)

If λ is larger, the data will move more closer to its nearest neighbor in the density field but might not in the distance field. The embedded data P(t) in the t-th iteration after density concentration process will be as follows.

$$P^{\left( t \right)} = P^{{\left( {t - 1} \right)}} + \lambda *\left( {P_{\text{NN}} - P^{{\left( {t - 1} \right)}} } \right)$$
(18)

3.3 DCMDS-RV algorithm

Although it is illustrated in our paper that LM method is capable of providing a better visualization than traditional MDS techniques, LM alone is not enough to generate impressive MDS results. Fortunately, by combining LM with the benefit of density concentration and NNM, the proposed DCMDS-RV algorithm can generate inspiring MDS results with data topology revealed. Detailed DCMDS-RV algorithm is presented in Fig. 2.

Fig. 2
figure 2

DCMDS-RV algorithm

3.4 Summary of DCMDS-RV algorithm

In summary, the proposed DCMDS-RV algorithm has several important merits as follows:

  1. 1.

    General MDS purpose DCMDS-RV algorithm applies LM method to find the embedded locations for data based on their mutual distances and therefore is a general-purpose MDS approach. Thus, it is suitable to conduct dimensionality reduction, visualization, and other purposes that general MDS approaches are used for. Besides, it is capable of escaping “colocated” problems that appear in traditional MDS techniques.

  2. 2.

    Micro-clusters forming Density-based clustering methodology (an unsupervised technique) is used to concentrate clusters so that the margins between micro-clusters are enlarged. As clusters are formed and separations between clusters expand, a better visualization of micro-clusters is expected.

  3. 3.

    Absence of parameter setting/integrated Simple and efficient, DCMDS-RV algorithm integrates cluster concentration based on local densities with MDS approach based on LM optimization by linear combination (19). It is easy to be interpreted and fulfilled. Most importantly, it is parameter-setting-free and unsupervised.

  4. 4.

    Revealing data topology NNMs are generated along the algorithm, which can provide data nearest neighbor information in density field. Unlike the traditional MDS approaches with only one purpose of embedding data, the proposed DCMDS-RV algorithm is capable of showing how the data are related/connected to the others. With the connections between related data shown, more information can be revealed in the embedded results.

4 Experimental results and analysis

In this section, the proposed DCMDS-RV algorithm is evaluated using experiments on four real-world data.

4.1 Human activity recognition (HAR) using smartphone dataset

HAR (Anguita et al. 2013) consists of 7352 data with 561 dimensions/attributes, built from the recordings of volunteers performing activities of daily living. These data belong to one of the following six classes: C1—walking, C2—walking upstairs, C3—walking downstairs, C4—sitting, C5—standing, and C6—laying.

In Fig. 3, Sammon mapping shows two distinct clusters with overlapping classes in each cluster. In Fig. 4, Isomap shows two distinct, dense clusters with overlapping classes of even less distinction compared to Fig. 3. LLE in Fig. 5 shows nearly no distinction among classes outside of being in the left or right loosely collected clusters. Similar to Sammon mapping, LAMP in Fig. 6 shows two distinct clusters with little distinction between classes within each cluster.

Fig. 3
figure 3

Sammon mapping on HAR

Fig. 4
figure 4

Isomap on HAR

Fig. 5
figure 5

LLE on HAR

Fig. 6
figure 6

LAMP on HAR

In contrast, DCMDS-RV in Fig. 7 shows distinct clusters with better class separation within each cluster. Moreover, the connections between data are clearly shown as well. Whereas some of the competing methods show sharp class distinction for no more than a few classes, DCMDS-RV shows distinct separation in almost every part of the map, save the two overlapping classes designated in pink and yellow. The optimal performance of DCMDS-RV for this dataset is verified in a later performance evaluation.

Fig. 7
figure 7

DCMDS-RV on HAR

4.2 MNIST handwritten digits dataset

MNIST handwritten digit recognitions have been considered as one of the most complex and difficult problems to be solved. It consists of 60,000 photographs with the size of 28 × 28 pixels (784 dimensions) (LeCun et al. 2017).

Figures 8, 9, 10, and 11 contain experiment results on first 6000 photographs of the MNIST dataset using Sammon mapping, Isomap, LLE, and LAMP, respectively. In Fig. 8, Sammon mapping shows that all data are contained in a large circle with many overlapping classes due to being trapped in local minima, demonstrating poor visualization; this technique is only able to distinguish a single class. Figure 9 shows that Isomap is able to distinguish the same class as Sammon mapping in a more compact, distinctive manner, yet struggles to show good visualization on the remaining overlapping classes. LLE visualization in Fig. 10 demonstrates better visualization than the former methods with a tight, distinct class clusters on the edges, but still has little distinction in the center. LAMP visualization also has small distinction in the edges, with most of the data barely distinguishable in the center. Unlike the other methods, the proposed DCMDS-RV in Fig. 12 shows distinct classes with significantly less overlap with visual separation between clusters. Additionally, the relations between data can be favorably visualized. Data relations given by NNM are critical to be analyzed in some cases, e.g., social media relations analysis (Cao et al. 2015). Similar to the HAR dataset, the proposed DCMDS-RV’s superiority is later verified in a performance evaluation.

Fig. 8
figure 8

Sammon mapping on MNIST

Fig. 9
figure 9

Isomap on MNIST

Fig. 10
figure 10

LLE on MNIST

Fig. 11
figure 11

LAMP on MNIST

Fig. 12
figure 12

DCMDS-RV on MNIST

4.3 Olivetti Face dataset

Face images (Du et al. 2018) are widely analyzed and applied to different applications. Here, Olivetti Face images are tested through our approach for visualization. Olivetti Face data (Samaria and Harter 1994) is a set of 112 × 92 (or 10,304 dimensions) images of different persons with different face angels, expressions, and even with or without glasses wearing. In Fig. 13, Sammon mapping shows relatively strong visualization results, with some overlapping classes seen throughout the map. Isomap, LLE, and LAMP in Figs. 14, 15, and 16 all fail to provide distinct class separation. In contrast, DCMDS-RV results in Fig. 17 provide significant class distinction with each datum collected to another, far outperforming all other methods. Only a single datum is incorrectly related. The Olivetti Face dataset clearly shows great performance of DCMDS-RV algorithm compared to the competing methods. Raw face images are eventually categorized into classes using the presented approach, as shown in Fig. 18.

Fig. 13
figure 13

Sammon mapping on Olivetti Faces

Fig. 14
figure 14

Isomap on Olivetti Faces

Fig. 15
figure 15

LLE on Olivetti Faces

Fig. 16
figure 16

LAMP on Olivetti Faces

Fig. 17
figure 17

DCMDS-RV on Olivetti Faces

Fig. 18
figure 18

Corresponding face images of Fig. 17

4.4 Wikipedia 2014 words dataset

Analysis on text such as text relation extraction reveals useful information. Here, Wikipedia corpuses are visualized using the proposed approach. Wikipedia corpuses (https://corpus.byu.edu/wiki/. accessed 24 January 2018) contain the full text of Wikipedia, and they contain 1.9 billion words in more than 4.4 million articles. Wikipedia 2014 words dataset is the Wikipedia corpuses of year 2014, with a vocabulary of the top 200,000 most frequent words. In the result of Fig. 19, the proposed approach provides an impressive visualization on text. The graph topology is successfully shown between texts. Four zoomed-in regions in the resulting graph are clearly illustrated in Fig. 20.

Fig. 19
figure 19

DCMDS-RV on Wikipedia 2014 words (956 samples)

Fig. 20
figure 20

Zoomed-in visualization results in the corresponding marked region in Fig. 19

4.5 Evaluation metrics

4.5.1 Kruskal stress

The concept of using a loss function to evaluate the performance of MDS came with Kruskal (1964) and gave us the concept of minimizing a loss function called stress. The disparity in stress is a measure of how well the Euclidean distance in low-dimensional space matches the dissimilarity, which is usually the Euclidean distance in high-dimensional space.

It’s proven in the original paper that the order of the original dissimilarities is preserved by the disparities. A loss function L, which is really stress, is defined as follows:

$$L = {\text{stress}} = \sqrt {\frac{{\mathop \sum \nolimits_{r < s} \left( {d_{rs} - \hat{d}_{rs} } \right)^{2} }}{{\mathop \sum \nolimits_{r < s} d_{rs}^{2} }}}$$
(20)

where drs is the Euclidean distance in low-dimensional space between point r and s. \(\hat{d}_{rs}\) is the disparity corresponding to drs. The order that indicates {r < s} is determined by the Euclidean distance relationships in high-dimensional space. When the MDS map perfectly reproduces the input data, \(d_{rs} - \hat{d}_{rs}\) is zero for all r and s, so stress is zero. Thus, generally speaking, the smaller the stress, the better the representation.

4.5.2 ACC

The standard unsupervised evaluation metric and protocols for evaluations and comparisons to other algorithms are used (Yang et al. 2010). Intuitively, this metric considers a cluster assignment from an unsupervised algorithm and a ground-truth assignment and then finds the best matching between them. The best mapping can be efficiently computed using the Hungarian algorithm (Kuhn 1955). For all the approaches, the number of clusters is set to be the number of ground-truth categories. Clustering performance is evaluated with unsupervised clustering accuracy (ACC):

$${\text{ACC}} = \mathop {\hbox{max} }\limits_{m} \frac{{\mathop \sum \nolimits_{i = 1}^{n} 1\left\{ {l_{i} = m\left( {c_{i} } \right)} \right\}}}{n}$$
(21)

where li is the ground-truth label, ci is the cluster assignment produced by the algorithm, and m ranges over all possible one-to-one mappings between clusters and labels.

4.6 Evaluation results and discussions

Kruskal stress factors using different MDS approaches for different datasets are shown in Table 1. Although other methods may have positive outcomes on some of the datasets, our method achieves better or same-level results each time.

Table 1 Comparison of Kruskal stress (s) and processing time (t in seconds): (s, t)

Experimental results (Table 2) using K-means (MacQueen 1967) to cluster embedded data clearly demonstrate a consistent superior performance of DCMDS-RV approach compared to other MDS methods.

Table 2 Comparison of ACC on four embedded datasets using K-means

As the iterative methods are not advanced in processing time, e.g., Sammon mapping and DCMDS-RV, our method has one limitation, which makes DCMDS-RV not favorable in fast mapping techniques. However, the visualization on data relations, in combination with the ability to produce category distinctions, favors DCMDS-RV approach as a highly successful data relation visualization method.

5 Conclusion

This paper proposes a new MDS algorithm called DCMDS-RV for data relation visualization in a low-dimensional space. The major innovation of the new algorithm is the incorporation of density concentration into our new MDS with LM optimization. DCMDS-RV algorithm can reveal both the distance relationships and the density relationships between data, and therefore successfully provide us a better visualization on all sorts of high-dimensional data as graph topology. Relations between data are revealed as NNM, showing how data are connected to the others. Experimental figures also give us vivid visualization results with relations shown. Compared with other state-of-the-art MDS methods, DCMDS-RV achieves better performance. Although it may suffer in embedding time, the proposed DCMDS-RV algorithm is a useful and heuristic technique for combining unsupervised clustering methods with traditional MDS techniques in further study.