1 Introduction

As one of the important technologies in computer vision, human posture estimation is a technology to obtain the current motion state of the target person through the image or the dynamic data of the target [1, 2]. It is widely used in games, identity recognition and sports performance analysis. At present, the biosensors used in the process of human posture estimation are difficult to capture the subtle posture changes of the human, so the Nano biosensor came into being. Nano biosensor is the fusion of nanotechnology and biosensor, using bio-specific recognition process to achieve the detection of sensor devices, bio-sensitive components including organisms, tissues, cells, organelles, cell membranes, enzymes, antibodies, nucleic acids and so on [3]. Its research involves many important fields such as biotechnology, information technology, nano science, interface science, and the integrated application of various advanced detection technologies such as light, sound, electricity and color, thus becoming the international research frontier and hotspot, compared with the general biosensor has a smaller size, faster, higher accuracy and reliability, free of labeling, good specificity, high efficiency and other advantages. The deformable convolutional neural networks (CNN) is an optimized form of convolutional neural network. Since the implementation of human posture estimation using CNN is complicated and the training time required is relatively long, the deformable convolutional neural network adds two modules of deformable convolution and deformable interest region pooling, which can be trained end-to-end by backward propagation algorithm to minimize the output error of the convolutional neural network and improve the training effect. Applying nano biosensor and deformable CNN to the human pose estimation process can improve the estimation accuracy and speed [4].

At present, many experts and scholars in various fields have conducted in-depth research into human posture estimation and put forward the human posture estimation algorithm. For example, Hong et al. [5] proposed an augmented human posture recognition algorithm. In this method, a training classifier is used to amplify the linear mechanical features of human motion, and a linear dynamic model is constructed. The human posture estimation results are obtained using this model. However, the effect of this algorithm in dealing with the noise of the moving target image is not good, resulting in the deviation of the estimated human posture. Qiao et al. [6] used human posture activation map as input, and used three-dimensional (3D) CNN to obtain human posture estimation results by training human posture joint data. In the application of this method, the selection of neural network parameters is subjective, and the model iteration is easy to fall into extreme value, so the application effect is poor. Singu et al. [7] proposed a human posture estimation method based on a single RGB image. Human posture estimation consists of two steps, namely, human posture feature extraction and posture point cloud estimation. CNN is used to extract and train data features. CNN can learn posture and feature data to recognize human posture. Further, the obtained features and the estimated human posture are used to obtain the point cloud of human posture. However, in practical application, it is found that the algorithm has the problem of low accuracy of human posture data. Song et al. [8] proposed a human pose estimation algorithm based on 3D multi-view basketball data set. After the RGB basketball motion pictures in the three-dimensional multi-view basketball motion data set are passed through the semantic segmentation network, the picture containing the target object is obtained, and the picture is input into the constructed feature fusion network model. After feature extraction of the RGB image and the depth image respectively, the RGB feature, the local feature of the point cloud and the global feature are stitched and fused to form a feature vector, and the human posture is recognized by combining the feature vector. However, in practical application, it is found that this algorithm is difficult to effectively recognize human postures such as leg lifting and squatting. Sengupta et al. [9] proposed a real-time human skeleton posture estimation algorithm based on millimeter wave radar and CNN. The reflected radar point cloud is analyzed and projected on the range azimuth and range elevation planes, and a low size high-resolution radar image representation method is designed, that is, the normalized values of the distance, elevation/azimuth and the power level of the reflected signal are allocated to the RGB channel. The algorithm uses a forked CNN structure and radar image representation to predict the real posture of bone joints in three-dimensional space, and realizes human posture estimation according to the prediction results. However, this algorithm has the problem of large estimation of human posture deviation.

To effectively address the problems of the above algorithms, a human posture estimation model is built through improved deep learning, and then the basic data of human motion posture and health is obtained using Nano biosensor acceleration sensors, angular velocity sensors in the data perception layer. The ability of human posture data can quickly and accurately estimate human posture. The main contributions of this paper are as follows: (1) The use of Nano biosensor to acquire human motion posture and health data solves the difficulty of traditional sensors to accurately capture subtle posture changes, and ensures the accuracy and speed of data collection results. (2) In this paper, the modified CNN model is improved and applied to human posture estimation, which solves the problem that the output result is not accurate if the parameter selection is unqualified. (3) The results using different data sets show that the proposed human posture estimation algorithm can effectively estimate the posture of human body such as leg lifting and squatting, and the application effect is good.

2 Methodology

2.1 Technical Framework of Human Pose Estimation Algorithm

The technical architecture of the human posture estimation algorithm is designed according to the layered idea, as shown in Fig. 1.

Fig. 1
figure 1

Technical architecture of human pose estimation algorithm

The technical architecture of human posture estimation algorithm is composed of data perception layer, service layer and an interaction layer. The basic data of human motion posture and health is obtained using nano-biological acceleration sensor, and angular velocity sensor in the data sensing layer. The nano-biological acceleration sensor is a kind of sensor that can measure acceleration. After the inertial force on the mass is measured, the acceleration value of the human is obtained using Newton’s second law of motion; the physical quantity measured by the angular velocity sensor is the rotational angular velocity when the human deflects and tilts; blood pressure sensor can ensure the smooth progress of accurate continuous measurement of human blood pressure. All sensors are arranged at appropriate postures to collect basic data on human motion posture and health. On this basis, the wireless sensor network is used to transmit all data to the data integration unit in the service layer. The protocol used in the wireless sensing network is a clustering-based routing protocol. The data integration unit converts the angular velocity and acceleration data in the basic human motion posture data to obtain the human motion posture data in the motion posture coordinate system.

Human motion posture and health data and physiological information recognition results are input. The deformed CNN model in the improved deep learning decision unit is used to estimate the human motion posture and health, and the estimation results are transmitted to the interaction layer. The human–machine interaction and service response unit in the interaction layer are used to present the human motion posture and health estimation results to the user. The structure of the Nano biosensor network is shown in Fig. 2.

Fig. 2
figure 2

Structure of nano biosensor network

According to the data in Fig. 2. In this paper, the nano-bio angular velocity sensor as well as the nano-bio acceleration sensor are installed in the monitoring area. The measurement range of the nano angular velocity sensor is ± 7200°, the sensitivity is mV/°, and the sampling frequency is 0–40 Hz, and the measurement range of the nano angular velocity sensor is ± 8 g, the sensitivity is 200 pm/g, and the sampling frequency is 0–40 Hz. After the Nano biological sensor obtains all the data in the process of human motion, it will send the data to the base station and transmit it to the task management node through the internet and satellite, so that it can be called at any time, ensuring the accuracy and efficiency of data collection.

2.2 Transformation of Motion Posture Coordinate System

After obtaining the angular velocity and acceleration of the human body in the process of motion using the nano-biological angular velocity sensor and the nano-biological acceleration sensor, it is necessary to convert the human body coordinate system and the geographical coordinate system where the angular velocity and acceleration are located to ensure the uniformity of the data [10, 11]. Here, the coordinate system in which the angular velocity and acceleration of Euler angle are located is used for conversion. The detailed process is as follows:

Let \(OX_{0} Y_{0} Z_{0}\) be human body coordinate system of angular velocity and acceleration. After the Euler angle is converted twice, the geographic coordinate system \(OXYZ\) of angular velocity and acceleration is obtained. The expression formula for converting human body coordinate system and geographical coordinate system is as follows:

$$\begin{gathered} OX_{0} Y_{0} Z_{0} \hfill \\ \mathop {}\limits^{{}} \mathop {}\limits^{{}} \downarrow X_{0} /\theta_{1} \hfill \\ OX_{1} Y_{1} Z_{1} \hfill \\ \mathop {}\limits^{{}} \mathop {}\limits^{{}} \downarrow Y_{1} /\theta_{2} \hfill \\ OX_{2} Y_{2} Z_{2} \hfill \\ \mathop {}\limits^{{}} \mathop {}\limits^{{}} \downarrow Z_{2} /\theta_{3} \hfill \\ OXYZ \hfill \\ \end{gathered}$$
(1)

where \(OX_{1} Y_{1} Z_{1}\) and \(OX_{2} Y_{2} Z_{2}\) refer to the first and second transition coordinate systems during coordinate conversion, respectively; \(\theta_{1}\), \(\theta_{2}\) and \(\theta_{3}\) refer to the rotation angle during coordinate system transformation, which is the Euler angle.

The angular relationship between the coordinate systems \(OX_{0} Y_{0} Z_{0}\) and \(OXYZ\) is described by converting the Euler angle into the coordinate system of angular velocity and acceleration. Then the direction cosine matrix formula of the coordinate system \(OX_{0} Y_{0} Z_{0}\) and the transition coordinate system at the first conversion is

$$G_{0}^{1} = \left[ {\begin{array}{*{20}l} 1 & 0 & 0 \\ 0 & {\cos \theta _{1} } & {\sin \theta _{1} } \\ 0 & { - \sin \theta _{1} } & {\cos \theta _{1} } \\ \end{array} } \right] \times \varepsilon ,$$
(2)

where \(G_{0}^{1}\) refers to the direction cosine matrix at the first conversion of coordinate system \(OX_{0} Y_{0} Z_{0}\); \(\varepsilon\) is the adjustment factor.

The direction cosine matrix \(G_{1}^{2}\) formula of the coordinate system \(OX_{0} Y_{0} Z_{0}\) and the transition coordinate system during the second conversion

$$G_{1}^{2} = \left[ {\begin{array}{*{20}l} {\cos \beta } & 0 & { - \sin \theta } \\ 0 & 1 & 0 \\ {\sin \theta _{2} } & 0 & {\cos \theta _{2} } \\ \end{array} } \right] \times \varepsilon ,$$
(3)

where \(\beta\) indicates the posture angle.

The expression formula of the direction cosine matrix between the transition coordinate system and the geographical coordinate system during the second conversion is as follows:

$$G_{1}^{2} = \left[ \begin{gathered} \cos \theta_{3} \mathop {}\limits^{{}} \mathop {}\limits^{{}} \sin \theta_{3} \mathop {}\limits^{{}} 0 \hfill \\ - \sin \theta_{3} \mathop {}\limits^{{}} \cos \theta_{3} \mathop {}\limits^{{}} 0 \hfill \\ \mathop {}\limits^{{}} \mathop {}\limits^{{}} 0\mathop {}\limits^{{}} \mathop {}\limits^{{}} \mathop {}\limits^{{}} \mathop {}\limits^{{}} \mathop {}\limits^{{}} 0\mathop {}\limits^{{}} \mathop {}\limits^{{}} 1 \hfill \\ \end{gathered} \right] \times \varepsilon .$$
(4)

According to the results of formulas (2)–(4), the conversion relationship between the angular velocity and acceleration coordinate system and the geographic coordinate system is obtained

$$\left[ \begin{gathered} x_{0} \hfill \\ y_{0} \hfill \\ x_{0} \hfill \\ \end{gathered} \right] = G\left[ \begin{gathered} x \hfill \\ y \hfill \\ z \hfill \\ \end{gathered} \right] \times \varepsilon ,$$
(5)

where \(x\), \(y\) and \(z\) refer to the three directions of the coordinate system, respectively; \(G\) refers to the direction cosine matrix of coordinate system transformation. And \(G = G_{2}^{3} G_{1}^{2} G_{0}^{1}\).

The transformation between the coordinate system of angular velocity and acceleration and the geographical coordinate system can be completed using formula (5).

2.3 Human Posture Estimation Model Based on Improved Deep Learning

CNN is the most widely used in deep learning. If the parameters of CNN are unqualified, the output results will be inaccurate. Therefore, a new deformable CNN is designed to improve the deep learning method.

The deformable CNN is derived from the CNN. The geometric deformation ability in the CNN model is added to make the output result of the CNN more accurate [12]. The angular velocity and acceleration coordinate conversion results are input into the front-end network of the deformed CNN model. After deformation bias processing and deformation convolution operation, the angular velocity, acceleration and other data are input into the self-inference network, and then a relay supervision module is constructed [13, 14]. The self-inference network is controlled by the module to output the human posture estimation results.

(1) Deformation bias of deformed CNN.

Let \(J_{{{\text{out}}}}\) be human posture features of the front-end network output of the deformable CNN. The expression formula is as follows:

$$J_{{{\text{out}}}} = {\text{Head(data),}}$$
(6)

where \({\text{Head}}( \cdot )\) refers to the mapping relation of front end of deformable CNN; \({\text{data}}\) refers to the angular velocity, acceleration and biological data of human body input into deformation CNN.

Geometric deformation processing is performed on the angular velocity, acceleration and human biological data in the front-end network of the deformation CNN, and the deformation bias is calculated first [15]. Set \(y\) as human posture features of the front-end network output of the deformable CNN. The feature point on the human posture feature is expressed by \(y(O_{0} )\). The expression formula of the feature point is

$$y(O_{0} ) = \sum\limits_{{p_{n} \in \hbar }} {w(o_{n} )} \times \sum\limits_{{p_{n} \in \hbar }} {\left[ {\left( {xo_{0} + xo_{n} } \right)} \right]} ,$$
(7)

where \(\hbar\) refers to the convolution kernel sampling grid; \(o_{n}\) denotes the position point corresponding to convolution kernel sampling network; \(O_{0}\) refers to the feature of human posture; \((xo_{0} + xo_{n} )\) refers to the weighted sum of human posture features and relative positions of convolution kernel;\(w(o_{n} )\) stands for the weight of human posture feature.

Set there are \(N\) relative position points in the deformed convolution mesh, the bias of the convolution kernel requires \(N \times 2\) bias outputs. When the human posture features are input into the deformation convolution kernel, the input channel and output channel are respectively \(C_{{{\text{in}}}}\) and \(C_{{{\text{out}}}}\), respectively. Then the bias value of the front-end network of the deformed CNN is \(H \times W \times C_{{{\text{in}}}} \times C_{{{\text{out}}}}\), in which \(W\) and \(H\) refer to the width and height values of human posture features.

Based on the bias value of the front-end network of the deformed CNN, the front-end network of the deformed CNN is deformed, as shown in Fig. 3.

Fig. 3
figure 3

Deformation process of deformation CNN

(2) Deformation convolution operation of deformation CNN.

After the basic data of human posture angular velocity and acceleration are input in the deformation CNN, the bias feature data is obtained through bias processing, and then the deformation convolution kernel operation is performed on the bias feature data, and the fused human posture data is obtained through back propagation[16, 17]. Then the self-inference network of the deformable CNN is constructed.

Set \(ZF_{0}\) as the output value of self-inference network of CNN. The expression formula is as follows:

$$ZF_{0} = {\text{SINet}}_{0} (J_{{{\text{out}}}} ),$$
(8)

where \({\text{SINet}}_{0} ( \cdot )\) represents the initial mapping relationship of the self-inference network.

When the series of self-inference networks is \(i\), its output numerical expression formula is as follows:

$$ZF_{i} = {\text{SINet}}_{i} \left( {\sum\limits_{j = 0, \ldots ,i - 1} {IS_{j} } } \right) + {\text{SINet}}_{i} (J_{{{\text{out}}}} ),$$
(9)

where \(IS_{j}\) represents the output value of the relay supervision module. \(j\) represents the number of relay supervision modules.

Set \(IS_{i}\) as the integrated human posture feature data output from the monitoring module of level \(i\), \(\varphi_{i}\) refers to the estimated thermal data of relay supervision module. Then, the relationship between \(IS_{i}\) and \(\varphi_{i}\) is expressed as:

$$IS_{i} ,\varphi_{i} = InS_{i} (ZF_{i} )$$
(10)

where \(InS_{i} ( \cdot )\) represents the mapping relationship of the level \(i\) relay supervision module.

The expression formula of \(\varphi_{i}\) is as follows:

$$\varphi_{N} = {\text{Fout}}(ZF_{N} ),$$
(11)

where \({\text{Fout}}( \cdot )\) refers to the mapping relation of final output results of deformed convolution network model; \(N\) refers to the number of relay supervision modules.

(3) Optimization of deformed CNN based on non-maximal value suppression.

Since the modified deformable CNN is easy to fall into the extreme value in the iterative process [18, 19], the parameterized posture non-maximum suppression method [20] is used to improve it.

Set \(d(U_{i} ,U_{j} \left| \wedge \right.)\) as the measurement function of human posture distance. This function is used to describe the similarity between human postures [21]. Then, the redundant posture elimination rule of \(d(U_{i} ,U_{j} \left| \wedge \right.)\) is as follows

$$d(U_{i} ,U_{j} \left| { \wedge ,\eta } \right.) = \left[ {d(U_{i} ,U_{j} \left| { \wedge ,\lambda } \right.) \times \delta \le \eta } \right],$$
(12)

where \(U_{i}\) and \(U_{j}\) refer to the \(i\) and \(i\) estimation results of human motion posture and health; \(d( \cdot )\) refers to the function, whose parameter is \(\wedge\); \(\eta\) and \(\lambda\) are the threshold and variable parameter; \(\delta\) is the constant number.

Set \(B_{i}\) as the target framework of human posture \(U_{i}\), and the matching function expression formula of the target box is

$$K_{{{\text{sim}}}} (U_{i} ,U_{j} \left| {\sigma_{1} } \right.) = \sum\limits_{n} {\tanh (c_{i}^{n} /\zeta_{1} ) \times \tanh (c_{j}^{n} /\zeta_{1} )} ,$$
(13)

where \(\zeta_{1}\) refers to the matching function parameter; \(\sigma_{1}\) and \(c_{j}\) refer to the credibility of human motion posture and health \(i\) and \(j\); \(n\) refers to the number of human motion posture and health.

Set the key feature distance of motion posture as \(H_{{{\text{sim}}}} (U_{i} ,U_{j} \left| {\zeta_{2} } \right.)\), and its expression formula is

$$H_{{{\text{sim}}}} (U_{i} ,U_{j} \left| {\zeta_{2} } \right.) = \sum\limits_{n} {\exp \left( { - \frac{{k_{i}^{n} }}{{\zeta_{2} }}} \right)} - \sum\limits_{n} {\exp \left( { - \frac{{k_{j}^{n} }}{{\zeta_{2} }}} \right)} ,$$
(14)

where \(\zeta_{2}\) refers to the distance function parameters of key features of motion posture; \(k_{i}^{n}\) and \(k_{j}^{n}\) refer to the key points of the \(i\) and \(j\) human body motion posture.

After formula (13) is combined with formula (14), the final human body estimated posture distance measurement function is obtained, and the expression formula is as follows:

$$d(U_{i} ,U_{j} \left| \wedge \right.) = \lambda \times \left[ {K_{{{\text{sim}}}} (U_{i} ,U_{j} \left| {\zeta_{1} } \right.) + H_{{{\text{sim}}}} (U_{i} ,U_{j} \left| \zeta \right._{2} )} \right].$$
(15)

After formula (15) is introduced into formula (8), the deformed CNN can avoid falling into the iterative extremum and realize human posture estimation. The human motion posture data are input into the optimized deformed CNN to obtain the relevant human posture estimation, which guarantees the maximum estimation accuracy and efficiency with good practical application-type results.

3 Experimental Results and Analysis

3.1 Data Sets

COCO dataset: like the ImageNet competition, it is regarded as one of the most concerned and authoritative competitions in the field of computer vision. After the suspension of the ImageNet competition, the COCO competition has become the most authoritative and important benchmark in the field of target recognition and detection. It is also the only competition in the world that can bring together Google, Microsoft, Facebook and many top universities and excellent innovative enterprises at home and abroad. The COCO dataset contains 200,000 images and more than 500,000 target annotations in 80 categories. It is a widely publicized target detection database, with an average number of targets per image of 7.2.

SURREAL dataset: The SURREAL dataset is a large-scale artificial gesture recognition dataset. For RGB video, it annotates a variety of states: depth information, body parts, optical flow, 3D gestures, etc. These images are the real rendering of people with great changes in shape, texture, viewpoint and posture, and contain 6 million frames of synthetic human body data.

The data in the two data sets were integrated, and 5000 images were randomly selected for experimental testing. The experimental data size was 23.56 G. In 5000 experimental images, 80% of the train sets and 20% were test sets. The data volume of the two sets is 18.71 G and 4.85 G, and the maximum number of features is 7. The data in the test set is input into the simulation software for trial operation. After several tests, the optimal experimental parameters were obtained, and the parameters were applied to the subsequent experiments, and the subsequent tests were completed using the data in the experimental set.

3.2 Evaluation Metrics

Based on the COCO dataset and the SURREAL data set, the motion posture is estimated using the algorithm in this paper, and the practical application effect of the algorithm is verified. At the same time, the algorithm in Variant grassmann manifolds [5], the algorithm in Lam-2srn [6], the algorithm in PCHP [7], the algorithm in HPRE [8], and the algorithm in mm-Pose [9] were used to experiment.

The human posture data acquisition accuracy is taken as an experimental index, the higher the value, the higher the accuracy of the obtained human posture data. The calculation formula of indicators is as follows:

$$A = \frac{{a_{2} }}{{a_{1} }} \times 100\% ,$$
(16)

where \(a_{1}\) is the amount of human posture data used in the experiment, and \(a_{2}\) is to the amount of human posture data accurately collected by different algorithms.

Fitting test: different methods have been used for fitting tests. The closer the fitting curve is to the feature point of human posture data, the higher the fitting degree and the better the practical application effect.

The calculation formula of motion posture estimation accuracy is as follows:

$$E = \frac{{D_{K} }}{{D_{Z} }} \times 100\% ,$$
(17)

where \(D_{K}\) represents the sample size of human posture correctly estimated by different methods.\(D_{Z}\) represents the total number of samples.

Human posture coincidence degree: this index refers to the coincidence degree between the human posture estimation result and the actual result under the number of jumps of the target. The calculation formula of this index is

$$W = \left| {w_{1} - w_{2} } \right|,$$
(18)

where \(w_{1}\) represents the estimation result of human posture. \(w_{2}\) is the actual result.

The estimation efficiency of this algorithm is tested when the number of estimated motion postures is different. The calculation formula of indicators is as follows:

$$T = \sum\limits_{i = 1}^{n} {t_{i} } ,$$
(19)

where \(t_{i}\) represents the time consumption of the \(i\)th human posture estimation step.\(n\) represents the total steps of human posture estimation.

3.3 Results and Discussion

The angular velocity, acceleration, blood pressure and other data obtained by the Nano biosensor are the basis of human posture recognition. Experimental data acquisition accuracy verifies the ability of this paper's algorithm to acquire human posture data using nano biosensor, which is the average accuracy of all nano biosensor, as shown in Table 1.

Table 1 Data acquisition accuracy (%)

According to the data in Table 1. The accuracy of human posture data acquisition of proposed algorithm shows a fluctuating trend, and reaches the maximum value of 99.6% at the 40th time. And the average accuracy of the human posture data acquisition of this algorithm is 98.1%, which is 11.5%, 5.2%, 1.5%, 21.5% and 7.9% higher than the algorithms in Variant Grassmann manifolds [5], Lam-2srn [6], PCHP [7], HPRE [8] and mm-Pose [9], respectively. The reason is that the algorithm uses nano- biosensors to acquire human posture data. Therefore, it ensures the accuracy and quality of data acquisition, which proves that the proposed algorithm has a strong ability to acquire human posture data.

Taking the fitting degree as the measurement index, the fitting degree of different algorithms is tested, and the results are shown in Fig. 4. It can be seen in Fig. 4.The fitting curve of the algorithm in Variant Grassmann manifolds [5] is close to the feature points of human posture data, but there is a large deviation in some areas. The fitting curve of the algorithm in Lam-2srn [6] coincides with the feature points of human posture data in the interval of [− 5, − 1] in the X direction, but there is a large distance from the feature points of human posture data. The algorithm in PCHP [7] is far from the feature point position of human posture data, while the algorithm in HPRE [8] has a large deviation from the feature point position of human posture data and poor fitting. The algorithm in mm-Pose [9] is far from the feature points of human posture data, while the trend of the fitting curve of the proposed algorithm when estimating human posture is completely the same as the distribution trend of the feature points of human posture data. Moreover, the fitting curve of the proposed algorithm is close to the position of the feature points of the human posture data, which indicates that there is no fitting phenomenon when estimating the human posture, and the output human posture estimation result is more accurate.

Fig. 4
figure 4

Comparison results of fitting

Table 2 shows the comparison results of motion posture estimation accuracy. The accuracy of motion posture estimation of this algorithm shows a fluctuating trend. When the number of experiments reaches 50, the maximum accuracy of motion posture estimation reaches 99.1%. The average estimation accuracy of this algorithm is 98.1%, which is 21.4%, 13.7%, 12.3%, 12.8% and 13% higher than the algorithms in Variant Grassmann manifolds [5], Lam-2srn [6], PCHP [7], HPRE [8] and mm-Pose [9], respectively. It shows that the proposed algorithm can effectively estimate human motion posture and health, and has certain application effect.

Table 2 Comparison results of motion attitude estimation accuracy (%)

Taking the coincidence degree of human posture recognition as a measurement index, the proposed algorithm is used to estimate the posture. At the same time, the algorithm in Variant grassmann manifolds [5], the algorithm in Lam-2srn [6], the algorithm in PCHP [7], the algorithm in HPRE [8], and the algorithm in mm-Pose [9] are also used to experiment, as shown in Fig. 5.

Fig. 5
figure 5

Comparison results of estimated human posture coincidence

According to the data in Fig. 5.The coincidence degree of the six human posture estimation methods for human posture recognition decreases with the increase of the number of jumps of the estimated target, and the decrease of the coincidence degree of the human posture estimated by the proposed algorithm, the algorithm in Lam-2srn [6] and the algorithm in HPRE [8] is similar. However, among the six methods, the proposed algorithm has the highest accuracy. When the estimated number of jumps of the target is 100 times, the coincidence degree of the target posture estimated by the proposed algorithm is about 98.5%, which is 3.5%, 1.2%, 5.1%, 1.9% and 6.1% higher than the algorithms in Variant Grassmann manifolds [5], Lam-2srn [6], PCHP [7], HPRE [8] and mm-Pose [9], respectively, indicating that the coincidence degree of the proposed algorithm is higher.

The estimation efficiency is taken as the performance index of the proposed algorithm, and it is tested under the different number of estimated motion postures. The results are shown in Fig. 6.

Fig. 6
figure 6

Comparison results of time efficiency for estimating human posture

It can be seen in Fig. 6. The time efficiency of the six algorithms for estimating human posture is inversely proportional to the amount of human posture data. Before the amount of human posture data is 2300, the time efficiency of this algorithm for estimating human posture shows a small upward trend. When the amount of human posture data exceeds 2300, the estimated human posture time of the proposed algorithm presents a large upward trend, lasting to 3000. When the amount of human posture data is between 3000 and 4800, the human posture estimation time presents a balanced state. When the amount of human posture data is between 4800 and 5600, the human posture estimation time shows a rapid upward trend. When the amount of human posture data is 10,000, the maximum time of human posture estimated by the proposed algorithm is only about 3.5 s, which is 7.8 s, 5.7 s, 4 s, 5.9 s and 8 s lower than the algorithms in Variant grassmann manifolds [5], Lam-2srn [6], PCHP [7], HPRE [8] and mm-Pose [9], respectively. The results show that the proposed algorithm has a short time to estimate human posture and is less affected by the amount of human posture data.

4 Conclusions

The paper applies Nano biosensor to the process of collecting human body estimation data. Nano biosensor has the characteristics of accurate and portable data acquisition. The experimental verification shows that the average accuracy of the human posture and health data acquisition of the proposed algorithm is 98.1%, and the trend of the fitting curve when estimating the human posture and health is the same as the distribution trend of the characteristic points of the human posture and health data, the average accuracy of the estimation is 98.1%, the coincidence of the estimated target posture is about 98.5%, and the maximum human posture and health estimation time is only about 3.5 s. Although the proposed algorithm has a strong ability in estimating human posture, there are still some technical defects in this paper. For example, the proposed algorithm does not de-noise the human posture and health data obtained by the Nano biosensor. Although the human posture and health data obtained by the Nano biosensor is relatively accurate, there are still different degrees of interference noise in the data, which affects the results of estimating human posture and health. Therefore, more research will be conducted on this in the future to solve the impact of noise on the estimation results.