Abstract
With the development of augmented reality supporting manual assembly collaboration, it is particularly important to explore the transformation from traditional “human-machine” cooperation mode to smart “human-machine” cooperation mode. Early studies have shown that user cues (i.e., head, gesture, eye) and scene cues (i.e., objects, tools, space) are intuitive and highly expressive for traditional AR collaborative mode. However, how to integrate these cues into an assembly system, reasonably infer an operator’s work intention, and then give an appropriate rendering scheme is one of the problems in collaborative assembly. This paper describes a AR collaborative assembly platform: SHARIDEAS. It uses a generalized grey correlation method to integrate user cues and scene cues. The results of data fusion can provide appropriate and intuitive assembly guidance for local workers. A formal user study is to explore the usability and feasibility of SHAREDEAS in a manual assembly task. The experimental data show that SHAREDEAS is more conducive than traditional one to improve the efficiency of human-machine cooperation. Finally, some conclusions of SHARIDEAS are given and the future research direction has prospected.
Similar content being viewed by others
Data availability
Not applicable.
Code availability
Not applicable.
References
Uva AE, Gattullo M, Manghisi VM, Spagnulo D, Cascella GL, Fiorentino M (2018) Evaluating the effectiveness of spatial augmented reality in smart manufacturing: a solution for manual working stations. Int J Adv Manuf Technol 94:509–521
Mengoni M, Ceccacci S, Generosi A, Leopardi A (2018) Spatial augmented reality: an application for human work in smart manufacturing environment. Procedia Manuf 17:476–483
Deshpande A, Kim I (2018) The effects of augmented reality on improving spatial problem solving for object assembly. Adv Eng Inform 38:760–775
Westerfield G, Mitrovic A, Billinghurst M (2015) Intelligent augmented reality training for motherboard assembly. Int J Artif Intell Educ 25:157–172
Zhou J, Lee I, Thomas B, Menassa R, Farrant A, Sansome A (2012) In-situ support for automotive manufacturing using spatial augmented reality. Int J Virtual Real 11:33–41
Kritzler M, Murr M, Michahelles F (2016) Remotebob: support of on-site workers via a telepresence remote expert system. In: Proceedings of the 6th International Conference on the Internet of Things, pp 7–14
Wang P, Zhang S, Bai X, Billinghurst M, He W, Sun M, Chen Y, Lv H, Ji H (2019) 2.5 DHANDS: a gesture-based MR remote collaborative platform. Int J Adv Manuf Technol 102:1339–1353
Wang P, Bai X, Billinghurst M, Zhang S, He W, Han D, Wang Y, Min H, Lan W, Han S (2020) Using a head pointer or eye gaze: the effect of gaze on spatial ar remote collaboration for physical tasks. Interact Comput 32:153–169
Gurevich P, Lanir J, Cohen B (2015) Design and implementation of teleadvisor: a projection-based augmented reality system for remote collaboration. Comput Support Cooper Work (CSCW) 24:527–562
Le Chénéchal M, Duval T, Gouranton V, Royan J, Arnaldi B (2016) Vishnu: virtual immersive support for HelpiNg users an interaction paradigm for collaborative remote guiding in mixed reality. In: 2016 IEEE Third VR International Workshop on Collaborative Virtual Environments (3DCVE), pp 9–12
Wang Z, Bai X, Zhang S, He W, Wang P, Zhang X et al (2020) SHARIdeas: a visual representation of intention sharing between designer and executor supporting AR assembly. In: SIGGRAPH Asia 2020 Posters, pp 1–2
Wang P, Zhang S, Billinghurst M, Bai X, He W, Wang S et al (2019) A comprehensive survey of AR/MR-based co-design in manufacturing. Eng Comput:1–24
Xuerui T, Julong D, Xiaojun C (2007) Generalized grey relational grade and grey relational order test, 2007 IEEE International Conference on Systems, Man and Cybernetics, pp 3928–3931
Piumsomboon T, Lee Y, Lee G, Billinghurst M (2017) In: SIGGRAPH Asia 2017 Emerging Technologies (ed) CoVAR: a collaborative virtual and augmented reality system for remote collaboration, pp 1–2
Piumsomboon T, Day A, Ens B, Lee Y, Lee G, Billinghurst M (2017) Exploring enhancements for remote mixed reality collaboration. In: SIGGRAPH Asia 2017 Mobile Graphics & Interactive Applications, pp 1–5
Wang P, Zhang S, Bai X, Billinghurst M, He W, Wang S et al (2019) Head pointer or eye gaze: which helps more in MR remote collaboration? In: 2019 IEEE conference on virtual reality and 3D user interfaces (VR), pp 1219–1220
Špakov O, Istance H, Räihä K-J, Viitanen T, Siirtola H (2019) Eye gaze and head gaze in collaborative games. In: Proceedings of the 11th ACM Symposium on Eye Tracking Research & Applications, pp 1–9
Renner P, Pfeiffer T (2017) Evaluation of attention guiding techniques for augmented reality-based assistance in picking and assembly tasks. In: Intelligent User Interfaces, pp 89–92
Erickson A, Norouzi N, Kim K, Schubert R, Jules J, LaViola JJ et al (2020) Sharing gaze rays for visual target identification tasks in collaborative augmented reality. J Multimodal User Interfaces 14:353–371
Norouzi N, Erickson A, Kim K, Schubert R, LaViola J, Bruder G et al (2019) Effects of shared gaze parameters on visual target identification task performance in augmented reality. In: Symposium on Spatial User Interaction, pp 1–11
Masai K, Kunze K, Sugimoto M, Billinghurst M (2016) Empathy glasses. In: Proceedings of the 2016 CHI Conference Extended Abstracts on Human Factors in Computing Systems, pp 1257–1263
Pan Y, Steed A (2016) Effects of 3D perspective on head gaze estimation with a multiview autostereoscopic display. Int J Human-Comput Stud 86:138–148
Steptoe W, Wolff R, Murgia A, Guimaraes E, Rae J, Sharkey P et al (2008) Eye-tracking for avatar eye-gaze and interactional analysis in immersive collaborative virtual environments. In: Proceedings of the 2008 ACM conference on Computer supported cooperative work, pp 197–200
Young J, Langlotz T, Cook M, Mills S, Regenbrecht H (2019) Immersive telepresence and remote collaboration using mobile and wearable devices. IEEE Trans Vis Comput Graph 25:1908–1918
Kim S, Lee G, Huang W, Kim H, Woo W, Billinghurst M (2019) Evaluating the combination of visual communication cues for HMD-based mixed reality remote collaboration. In: Proceedings of the 2019 CHI conference on human factors in computing systems, pp 1–13
Bai H, Sasikumar P, Yang J, Billinghurst M (2020) A user study on mixed reality remote collaboration with eye gaze and hand gesture sharing. In: Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp 1–13
Gao L, Bai H, Lindeman R, Billinghurst M (2017) Static local environment capturing and sharing for MR remote collaboration. In: SIGGRAPH Asia 2017 Mobile Graphics & Interactive Applications, pp 1–6
Gao L, Bai H, He W, Billinghurst M, Lindeman RW (2018) Real-time visual representations for mobile mixed reality remote collaboration. In: SIGGRAPH Asia 2018 Virtual & Augmented Reality, pp 1–2
Lee GA, Teo T, Kim S, Billinghurst M (2017) Mixed reality collaboration through sharing a live panorama. In: SIGGRAPH Asia 2017 Mobile Graphics & Interactive Applications, pp 1–4
Piumsomboon T, Lee GA, Irlitti A, Ens B, Thomas BH, Billinghurst M (2019) On the shoulder of the giant: A multi-scale mixed reality collaboration with 360 video sharing and tangible interaction. In: Proceedings of the 2019 CHI conference on human factors in computing systems, pp 1–17
Wang P, Bai X, Billinghurst M, Zhang S, Wei S, Xu G et al (2020) 3DGAM: using 3D gesture and CAD models for training on mixed reality remote collaboration. Multimed Tools Appl:1–26
Wang Y, Zhang S, Yang S, He W, Bai X (2018) Mechanical assembly assistance using marker-less augmented reality system. Assem Autom
Wang Y, Zhang S, Wan B, He W, Bai X (2018) Point cloud and visual feature-based tracking method for an augmented reality-aided mechanical assembly system. Int J Adv Manuf Technol 99:2341–2352
Wang Z, Bai X, Zhang S, He W, Zhang X, Zhang L, Wang P, Han D, Yan Y (2020) Information-level AR instruction: a novel assembly guidance information representation assisting user cognition. Int J Adv Manuf Technol 106:603–626
Wang Z, Bai X, Zhang S, He W, Zhang X, Yan Y et al (2020) Information-level real-time AR instruction: a novel dynamic assembly guidance information representation assisting human cognition. Int J Adv Manuf Technol:1–19
M. Bose, "Optimal crossover designs," in International Conference on Parallel Processing Workshops, 2009.
Gavish N, Gutierrez T, Webel S, Rodriguez J, Peveri M, Bockholt U et al (2015) Evaluating virtual reality and augmented reality training for industrial maintenance and assembly tasks. Interact Learn Environ 23:778–798
Houpt JW, Blaha LM (2016) Mathematical models of perception and cognition volume I: A Festschrift for James T. Psychology Press, Townsend
Henderson SJ, Feiner SK (2011) Augmented reality in the psychomotor phase of a procedural task. In: 2011 10th IEEE International Symposium on Mixed and Augmented Reality, pp 191–19s
Acknowledgments
Thanks to Dong Han of the University of Toronto for checking the English manuscript of the earlier version and he helped authors correct the grammatical errors in the paper.
Funding
This work is partly supported by Seed Foundation of Innovation and Creation for Graduate Students in Northwestern Polytechnical University (ZZ2018084), Civil Aircraft Special Project (MJZ-2017-G73), and SASTIND China under Grant (JCKY2018205B021).
Author information
Authors and Affiliations
Contributions
SF and XH provided some valuable design solutions for this UX experiment. JZ and YZ established the basic hardware environment for our research. YW broke through the technical difficulty of this research, and JZ did a lot of work for the collection of experimental data. In particular, we would like to thank APXB and Prof. SZ for their constructive comments on the improvement of the experiment.
Corresponding author
Ethics declarations
Competing interests
Our team declares that we have no financial and personal relationships with other people or organizations that can inappropriately influence our work; there is no professional or other personal interest of any nature or kind in any product, service, and/or company that could be construed as influencing the position presented in, or the review of, the manuscript entitled.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Definition of Generalized Grey Relation
If the behavior factor of SHARIDEAS is x0 and x0 is affected by multiple factors xi (i=1, 2,…, n), the method of expressing xi's influence on x0 by the grey correlation degree of factor xi and factor x0 is called generalized grey correlation. Based on this concept, we regard the above visual cues as a grey system, and the evaluation of the fusion process is determined by various indicators.
Taking the fusion process of visual cues as a research object, the system characteristic behavior sequence of operator Intention Evaluation Index x0 is as follows:
In Formula 1, x0(k) is the kth evaluation data of a behavior factor in the process of intention perception.
According to the cognitive requirements of various visual cues, the quality of intention cognition depends on the fusion of various behavioral factors. On this basis, the grey relational mathematical model is constructed.
Assuming that there are m factors influencing intention calculation, the effective date of relevant visual cues are extracted by sensors (e.g., industrial camera), and the behavioral index sequence of the ith influencing factor xi is as follows:
In formula 2, i is the number of the ith influencing factor(i = 1, 2,…, m); k is the serial number of the data (k = 1, 2,…, n); x1(k)is the kth data collected by influencing factor xi.
From Formula 1 and 2, the corresponding forms of system characteristic behavior sequencex0 and comparative data sequence xi are deduced, then:
Fusion parameter calculation of visual cues
In collaborative assembly, visual cues can be divided into user cues (UC) and environment cues (EC). User cues include gesture cues, head gaze cues, and eye gaze cues, while environment cues mainly include spatial orientation cues, tool cues, and assembly objects.
The spatial relationship between the components of the second stage reducer is complex. According to the requirements of its working conditions and technical parameters, such as the assembly of the upper and lower box, the center distance of the box, the dividing circle of the gear shaft, the airtightness requirements of the internal environment, the vibration resistance, and impact resistance, etc., 10 evaluation coefficients of visual cues are determined.
Due to the different numerical units of the two types of visual cues, SHARIDEAS integrates the collected visual cues into a multi-dimensional evaluation. Its evaluation results are divided into 10 grades: when the evaluation coefficient is 8–10, it is an excellent factor; when the evaluation coefficient is 6–7, it is a good factor; when the evaluation coefficient is 4–5, it is determined as a general factor; when the evaluation coefficient is 1–3, it is a general factor, that is, a secondary factor. Therefore, according to the current intention of user, the 10 indexes will be sorted adaptively. Thus, the system characteristic behavior sequence X0 composed of 10 evaluation coefficients is obtained. In the process of collaborative assembly, SHARIDEAS calculates the corresponding comparative data sequence Xi according to the real-time data.
Then, the interval value is processed according to following formulas.
On this basis, the data series \( {X}_0^{\prime } \) and \( {X}_i^{\prime } \) after interval value processing are obtained. The data are further processed according to the following formula:
For each group of sequences, the maximum and minimum values are obtained.
If the resolution coefficient ζ∈ (0,1) is defined, then,
According to formula (8) ~ Eq. (11), SHAREIDEAS performs the following operations:
γ0i’s data itself represents the intention tendency of the clue. For example, in the input cues, the evaluation parameters of the “twist” gesture range from 0.67 to 0.82. When γ0i’s data is 0.75, this cue dominates the user’s intention.
Rights and permissions
About this article
Cite this article
Wang, Z., Wang, Y., Bai, X. et al. SHARIDEAS: a smart collaborative assembly platform based on augmented reality supporting assembly intention recognition. Int J Adv Manuf Technol 115, 475–486 (2021). https://doi.org/10.1007/s00170-021-07142-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00170-021-07142-y