Skip to main content
Log in

SHARIDEAS: a smart collaborative assembly platform based on augmented reality supporting assembly intention recognition

  • ORIGINAL ARTICLE
  • Published:
The International Journal of Advanced Manufacturing Technology Aims and scope Submit manuscript

Abstract

With the development of augmented reality supporting manual assembly collaboration, it is particularly important to explore the transformation from traditional “human-machine” cooperation mode to smart “human-machine” cooperation mode. Early studies have shown that user cues (i.e., head, gesture, eye) and scene cues (i.e., objects, tools, space) are intuitive and highly expressive for traditional AR collaborative mode. However, how to integrate these cues into an assembly system, reasonably infer an operator’s work intention, and then give an appropriate rendering scheme is one of the problems in collaborative assembly. This paper describes a AR collaborative assembly platform: SHARIDEAS. It uses a generalized grey correlation method to integrate user cues and scene cues. The results of data fusion can provide appropriate and intuitive assembly guidance for local workers. A formal user study is to explore the usability and feasibility of SHAREDEAS in a manual assembly task. The experimental data show that SHAREDEAS is more conducive than traditional one to improve the efficiency of human-machine cooperation. Finally, some conclusions of SHARIDEAS are given and the future research direction has prospected.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Data availability

Not applicable.

Code availability

Not applicable.

References

  1. Uva AE, Gattullo M, Manghisi VM, Spagnulo D, Cascella GL, Fiorentino M (2018) Evaluating the effectiveness of spatial augmented reality in smart manufacturing: a solution for manual working stations. Int J Adv Manuf Technol 94:509–521

    Article  Google Scholar 

  2. Mengoni M, Ceccacci S, Generosi A, Leopardi A (2018) Spatial augmented reality: an application for human work in smart manufacturing environment. Procedia Manuf 17:476–483

    Article  Google Scholar 

  3. Deshpande A, Kim I (2018) The effects of augmented reality on improving spatial problem solving for object assembly. Adv Eng Inform 38:760–775

    Article  Google Scholar 

  4. Westerfield G, Mitrovic A, Billinghurst M (2015) Intelligent augmented reality training for motherboard assembly. Int J Artif Intell Educ 25:157–172

    Article  Google Scholar 

  5. Zhou J, Lee I, Thomas B, Menassa R, Farrant A, Sansome A (2012) In-situ support for automotive manufacturing using spatial augmented reality. Int J Virtual Real 11:33–41

    Article  Google Scholar 

  6. Kritzler M, Murr M, Michahelles F (2016) Remotebob: support of on-site workers via a telepresence remote expert system. In: Proceedings of the 6th International Conference on the Internet of Things, pp 7–14

    Chapter  Google Scholar 

  7. Wang P, Zhang S, Bai X, Billinghurst M, He W, Sun M, Chen Y, Lv H, Ji H (2019) 2.5 DHANDS: a gesture-based MR remote collaborative platform. Int J Adv Manuf Technol 102:1339–1353

    Article  Google Scholar 

  8. Wang P, Bai X, Billinghurst M, Zhang S, He W, Han D, Wang Y, Min H, Lan W, Han S (2020) Using a head pointer or eye gaze: the effect of gaze on spatial ar remote collaboration for physical tasks. Interact Comput 32:153–169

    Article  Google Scholar 

  9. Gurevich P, Lanir J, Cohen B (2015) Design and implementation of teleadvisor: a projection-based augmented reality system for remote collaboration. Comput Support Cooper Work (CSCW) 24:527–562

    Article  Google Scholar 

  10. Le Chénéchal M, Duval T, Gouranton V, Royan J, Arnaldi B (2016) Vishnu: virtual immersive support for HelpiNg users an interaction paradigm for collaborative remote guiding in mixed reality. In: 2016 IEEE Third VR International Workshop on Collaborative Virtual Environments (3DCVE), pp 9–12

    Chapter  Google Scholar 

  11. Wang Z, Bai X, Zhang S, He W, Wang P, Zhang X et al (2020) SHARIdeas: a visual representation of intention sharing between designer and executor supporting AR assembly. In: SIGGRAPH Asia 2020 Posters, pp 1–2

    Google Scholar 

  12. Wang P, Zhang S, Billinghurst M, Bai X, He W, Wang S et al (2019) A comprehensive survey of AR/MR-based co-design in manufacturing. Eng Comput:1–24

  13. Xuerui T, Julong D, Xiaojun C (2007) Generalized grey relational grade and grey relational order test, 2007 IEEE International Conference on Systems, Man and Cybernetics, pp 3928–3931

    Google Scholar 

  14. Piumsomboon T, Lee Y, Lee G, Billinghurst M (2017) In: SIGGRAPH Asia 2017 Emerging Technologies (ed) CoVAR: a collaborative virtual and augmented reality system for remote collaboration, pp 1–2

  15. Piumsomboon T, Day A, Ens B, Lee Y, Lee G, Billinghurst M (2017) Exploring enhancements for remote mixed reality collaboration. In: SIGGRAPH Asia 2017 Mobile Graphics & Interactive Applications, pp 1–5

    Google Scholar 

  16. Wang P, Zhang S, Bai X, Billinghurst M, He W, Wang S et al (2019) Head pointer or eye gaze: which helps more in MR remote collaboration? In: 2019 IEEE conference on virtual reality and 3D user interfaces (VR), pp 1219–1220

    Chapter  Google Scholar 

  17. Špakov O, Istance H, Räihä K-J, Viitanen T, Siirtola H (2019) Eye gaze and head gaze in collaborative games. In: Proceedings of the 11th ACM Symposium on Eye Tracking Research & Applications, pp 1–9

    Google Scholar 

  18. Renner P, Pfeiffer T (2017) Evaluation of attention guiding techniques for augmented reality-based assistance in picking and assembly tasks. In: Intelligent User Interfaces, pp 89–92

    Google Scholar 

  19. Erickson A, Norouzi N, Kim K, Schubert R, Jules J, LaViola JJ et al (2020) Sharing gaze rays for visual target identification tasks in collaborative augmented reality. J Multimodal User Interfaces 14:353–371

    Article  Google Scholar 

  20. Norouzi N, Erickson A, Kim K, Schubert R, LaViola J, Bruder G et al (2019) Effects of shared gaze parameters on visual target identification task performance in augmented reality. In: Symposium on Spatial User Interaction, pp 1–11

    Google Scholar 

  21. Masai K, Kunze K, Sugimoto M, Billinghurst M (2016) Empathy glasses. In: Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, pp 1257–1263

    Chapter  Google Scholar 

  22. Pan Y, Steed A (2016) Effects of 3D perspective on head gaze estimation with a multiview autostereoscopic display. Int J Human-Comput Stud 86:138–148

    Article  Google Scholar 

  23. Steptoe W, Wolff R, Murgia A, Guimaraes E, Rae J, Sharkey P et al (2008) Eye-tracking for avatar eye-gaze and interactional analysis in immersive collaborative virtual environments. In: Proceedings of the 2008 ACM conference on Computer supported cooperative work, pp 197–200

    Chapter  Google Scholar 

  24. Young J, Langlotz T, Cook M, Mills S, Regenbrecht H (2019) Immersive telepresence and remote collaboration using mobile and wearable devices. IEEE Trans Vis Comput Graph 25:1908–1918

    Article  Google Scholar 

  25. Kim S, Lee G, Huang W, Kim H, Woo W, Billinghurst M (2019) Evaluating the combination of visual communication cues for HMD-based mixed reality remote collaboration. In: Proceedings of the 2019 CHI conference on human factors in computing systems, pp 1–13

    Google Scholar 

  26. Bai H, Sasikumar P, Yang J, Billinghurst M (2020) A user study on mixed reality remote collaboration with eye gaze and hand gesture sharing. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp 1–13

    Google Scholar 

  27. Gao L, Bai H, Lindeman R, Billinghurst M (2017) Static local environment capturing and sharing for MR remote collaboration. In: SIGGRAPH Asia 2017 Mobile Graphics & Interactive Applications, pp 1–6

    Google Scholar 

  28. Gao L, Bai H, He W, Billinghurst M, Lindeman RW (2018) Real-time visual representations for mobile mixed reality remote collaboration. In: SIGGRAPH Asia 2018 Virtual & Augmented Reality, pp 1–2

    Google Scholar 

  29. Lee GA, Teo T, Kim S, Billinghurst M (2017) Mixed reality collaboration through sharing a live panorama. In: SIGGRAPH Asia 2017 Mobile Graphics & Interactive Applications, pp 1–4

    Google Scholar 

  30. Piumsomboon T, Lee GA, Irlitti A, Ens B, Thomas BH, Billinghurst M (2019) On the shoulder of the giant: A multi-scale mixed reality collaboration with 360 video sharing and tangible interaction. In: Proceedings of the 2019 CHI conference on human factors in computing systems, pp 1–17

    Google Scholar 

  31. Wang P, Bai X, Billinghurst M, Zhang S, Wei S, Xu G et al (2020) 3DGAM: using 3D gesture and CAD models for training on mixed reality remote collaboration. Multimed Tools Appl:1–26

  32. Wang Y, Zhang S, Yang S, He W, Bai X (2018) Mechanical assembly assistance using marker-less augmented reality system. Assem Autom

  33. Wang Y, Zhang S, Wan B, He W, Bai X (2018) Point cloud and visual feature-based tracking method for an augmented reality-aided mechanical assembly system. Int J Adv Manuf Technol 99:2341–2352

    Article  Google Scholar 

  34. Wang Z, Bai X, Zhang S, He W, Zhang X, Zhang L, Wang P, Han D, Yan Y (2020) Information-level AR instruction: a novel assembly guidance information representation assisting user cognition. Int J Adv Manuf Technol 106:603–626

    Article  Google Scholar 

  35. Wang Z, Bai X, Zhang S, He W, Zhang X, Yan Y et al (2020) Information-level real-time AR instruction: a novel dynamic assembly guidance information representation assisting human cognition. Int J Adv Manuf Technol:1–19

  36. M. Bose, "Optimal crossover designs," in International Conference on Parallel Processing Workshops, 2009.

    Google Scholar 

  37. Gavish N, Gutierrez T, Webel S, Rodriguez J, Peveri M, Bockholt U et al (2015) Evaluating virtual reality and augmented reality training for industrial maintenance and assembly tasks. Interact Learn Environ 23:778–798

    Article  Google Scholar 

  38. Houpt JW, Blaha LM (2016) Mathematical models of perception and cognition volume I: A Festschrift for James T. Psychology Press, Townsend

    Book  Google Scholar 

  39. Henderson SJ, Feiner SK (2011) Augmented reality in the psychomotor phase of a procedural task. In: 2011 10th IEEE International Symposium on Mixed and Augmented Reality, pp 191–19s

    Chapter  Google Scholar 

Download references

Acknowledgments

Thanks to Dong Han of the University of Toronto for checking the English manuscript of the earlier version and he helped authors correct the grammatical errors in the paper.

Funding

This work is partly supported by Seed Foundation of Innovation and Creation for Graduate Students in Northwestern Polytechnical University (ZZ2018084), Civil Aircraft Special Project (MJZ-2017-G73), and SASTIND China under Grant (JCKY2018205B021).

Author information

Authors and Affiliations

Authors

Contributions

SF and XH provided some valuable design solutions for this UX experiment. JZ and YZ established the basic hardware environment for our research. YW broke through the technical difficulty of this research, and JZ did a lot of work for the collection of experimental data. In particular, we would like to thank APXB and Prof. SZ for their constructive comments on the improvement of the experiment.

Corresponding author

Correspondence to Xiaoliang Bai.

Ethics declarations

Competing interests

Our team declares that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work; there is no professional or other personal interest of any nature or kind in any product, service, and/or company that could be construed as influencing the position presented in, or the review of, the manuscript entitled.

Additional information

Publisher’s note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix

Appendix

Definition of Generalized Grey Relation

If the behavior factor of SHARIDEAS is x0 and x0 is affected by multiple factors xi (i=1, 2,…, n), the method of expressing xi's influence on x0 by the grey correlation degree of factor xi and factor x0 is called generalized grey correlation. Based on this concept, we regard the above visual cues as a grey system, and the evaluation of the fusion process is determined by various indicators.

Taking the fusion process of visual cues as a research object, the system characteristic behavior sequence of operator Intention Evaluation Index x0 is as follows:

$$ {X}_0=\left\{{x}_0(1),{x}_0(2),\dots, {x}_0(k),\dots, {x}_0(n)\right\} $$
(1)

In Formula 1, x0(k) is the kth evaluation data of a behavior factor in the process of intention perception.

According to the cognitive requirements of various visual cues, the quality of intention cognition depends on the fusion of various behavioral factors. On this basis, the grey relational mathematical model is constructed.

Assuming that there are m factors influencing intention calculation, the effective date of relevant visual cues are extracted by sensors (e.g., industrial camera), and the behavioral index sequence of the ith influencing factor xi is as follows:

$$ {X}_i=\left\{{x}_i(1),{x}_i(2),\dots, {x}_i(k),\dots, {x}_i(n)\right\} $$
(2)

In formula 2, i is the number of the ith influencing factor(i = 1, 2,…, m); k is the serial number of the data (k = 1, 2,…, n); x1(k)is the kth data collected by influencing factor xi.

From Formula 1 and 2, the corresponding forms of system characteristic behavior sequencex0 and comparative data sequence xi are deduced, then:

$$ {\displaystyle \begin{array}{l}{X}_0=\left\{{x}_0(1),{x}_0(2),\dots, {x}_0(k),\dots, {x}_0(n)\right\}\\ {}\kern3em \vdots \\ {}{X}_i=\left\{{x}_i(1),{x}_i(2),\dots, {x}_i(k),\dots, {x}_i(n)\right\}\\ {}\kern3em \vdots \\ {}{X}_m=\left\{{x}_m(1),{x}_m(2),\dots, {x}_m(k),\dots, {x}_m(n)\right\}\end{array}} $$
(3)

Fusion parameter calculation of visual cues

In collaborative assembly, visual cues can be divided into user cues (UC) and environment cues (EC). User cues include gesture cues, head gaze cues, and eye gaze cues, while environment cues mainly include spatial orientation cues, tool cues, and assembly objects.

The spatial relationship between the components of the second stage reducer is complex. According to the requirements of its working conditions and technical parameters, such as the assembly of the upper and lower box, the center distance of the box, the dividing circle of the gear shaft, the airtightness requirements of the internal environment, the vibration resistance, and impact resistance, etc., 10 evaluation coefficients of visual cues are determined.

Due to the different numerical units of the two types of visual cues, SHARIDEAS integrates the collected visual cues into a multi-dimensional evaluation. Its evaluation results are divided into 10 grades: when the evaluation coefficient is 8–10, it is an excellent factor; when the evaluation coefficient is 6–7, it is a good factor; when the evaluation coefficient is 4–5, it is determined as a general factor; when the evaluation coefficient is 1–3, it is a general factor, that is, a secondary factor. Therefore, according to the current intention of user, the 10 indexes will be sorted adaptively. Thus, the system characteristic behavior sequence X0 composed of 10 evaluation coefficients is obtained. In the process of collaborative assembly, SHARIDEAS calculates the corresponding comparative data sequence Xi according to the real-time data.

Then, the interval value is processed according to following formulas.

$$ {X}_iB=\left\{{x}_i(1)b,{x}_i(2)b,\dots, {x}_i(k)b,\dots, {x}_i(n)b\right\} $$
(4)
$$ {x}_i(k)b=\frac{x_i(k)-\min {x}_i(k)}{\max {x}_i(k)-\min {x}_i(k)},\kern0.5em k=1,2,\dots, n $$
(5)

On this basis, the data series \( {X}_0^{\prime } \) and \( {X}_i^{\prime } \) after interval value processing are obtained. The data are further processed according to the following formula:

$$ {\Delta}_i(k)=\left|{x}_0^{\prime }(k)-{x}_i^{\prime }(k)\right| $$
(6)
$$ {\Delta}_i=\left|{\Delta}_i(1),{\Delta}_i(2),\dots, {\Delta}_i(n)\right|i=0,1,\dots, m $$
(7)

For each group of sequences, the maximum and minimum values are obtained.

$$ M=\underset{i}{\mathit{\max}}\underset{k}{\max }{\Delta}_i(k) $$
(8)
$$ m=\underset{i}{\mathit{\min}}\underset{k}{\mathit{\min}}{\Delta}_i(k) $$
(9)

If the resolution coefficient ζ∈ (0,1) is defined, then,

$$ \gamma \left({x}_0(k),{x}_i(k)\right)=\frac{\underset{i}{\min}\underset{k}{\min}\left|{x}_0(k)-{x}_i(k)\right|+\zeta \underset{i}{\max}\underset{k}{\max}\left|{x}_0(k)-{x}_i(k)\right|}{\left|{x}_0(k)-{x}_i(k)\right|+\zeta \underset{i}{\max}\underset{k}{\max}\left|{x}_0(k)-{x}_i(k)\right|} $$
(10)
$$ \gamma \left({X}_0,{X}_i\right)=\frac{1}{n}\sum \limits_{k=1}^n\gamma \left({x}_0(k),{x}_i(k)\right) $$
(11)

According to formula (8) ~ Eq. (11), SHAREIDEAS performs the following operations:

$$ {\gamma}_{0i}(k)=\frac{m+\zeta M}{\Delta_i(k)-\zeta M} $$
(12)
$$ {\beta}_{0i}(k)=\left\{{\gamma}_{0i}(1),{\gamma}_{0i}(2),\dots, {\gamma}_{0i}(k),\dots, {\gamma}_{0i}(n)\right\} $$
(13)
$$ {\gamma}_{0i}=\frac{1}{n}\sum \limits_{k=1}^n{\gamma}_{0i}(k) $$
(14)

γ0i’s data itself represents the intention tendency of the clue. For example, in the input cues, the evaluation parameters of the “twist” gesture range from 0.67 to 0.82. When γ0i’s data is 0.75, this cue dominates the user’s intention.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wang, Z., Wang, Y., Bai, X. et al. SHARIDEAS: a smart collaborative assembly platform based on augmented reality supporting assembly intention recognition. Int J Adv Manuf Technol 115, 475–486 (2021). https://doi.org/10.1007/s00170-021-07142-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00170-021-07142-y

Keywords

Navigation