Skip to main content

Understanding Shared Autonomy of Collaborative Humans Using Motion Capture System for Simulating Team Assembly

  • Conference paper
  • First Online:
Towards Sustainable Customization: Bridging Smart Products and Manufacturing Systems (CARV 2021, MCPC 2021)

Abstract

In virtual production planning, simulating human motions helps to improve process planning and interaction efficiency. However, simulating multiple humans sharing tasks in a shared workplace requires understanding how human workers interact and share autonomy. In this regard, an Inertial Measurement Unit based motion capture is employed for understanding shifting roles and learning effects. Parameters such as total time, distance, and acceleration variances in repetition are considered for modeling collaborative motion interactions. The results distinguish motion patterns versus the undertaken interactions. This work may serve as an initial input to model interaction schemes and recognize human actions behavior during team assembly. Furthermore, the concept can be extended toward a human-robot shared autonomy.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 299.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 379.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 379.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Billard, A., Calinon, S., Dillmann, R., Schaal, S.: Robot Programming by Demonstration. In: Siciliano, B., Khatib, O. (eds.) Springer Handbook of Robotics, pp. 1371–1394. Springer, Heidelberg (2008). https://doi.org/10.1007/978-3-540-30301-5_60

  2. Blake, R., Shiffrar, M.: Perception of human motion. Ann. Rev. Psychol. 58(1), 47–73 (2006). https://www.annualreviews.org/doi/10.1146/annurev.psych.57.102904.190152

  3. Carrara, F., Elias, P., Sedmidubsky, J., Zezula, P.: LSTM-based real-time action detection and prediction in human motion streams. Multimedia Tools Appl. 78(19), 27309–27331 (2019). https://doi.org/10.1007/s11042-019-07827-3

  4. Elhayek, A., Kovalenko, O., Murthy, P., Malik, J., Stricker, D.: Fully automatic multi-person human motion capture for VR applications. In: Bourdot, P., Cobb, S., Interrante, V., kato, H., Stricker, D. (eds.) Virtual Reality and Augmented Reality, pp. 28–47. Lecture Notes in Computer Science, Springer, Cham (2018). https://doi.org/10.1007/978-3-030-01790-3_3

  5. Filippeschi, A., Schmitz, N., Miezal, M., Bleser, G., Ruffaldi, E., Stricker, D.: Survey of motion tracking methods based on inertial sensors: a focus on upper limb human motion. Sensors (Basel, Switzerland) 17(6) (2017). https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5492902/

  6. Jekel, C.F., Venter, G., Venter, M.P., Stander, N., Haftka, R.T.: Similarity measures for identifying material parameters from hysteresis loops using inverse analysis. Int. J. Mater. Form. (2019). https://doi.org/10.1007/s12289-018-1421-8

  7. Liu, Z., Liu, Q., Xu, W., Liu, Z., Zhou, Z., Chen, J.: Deep learning-based human motion prediction considering context awareness for human-robot collaboration in manufacturing. Procedia CIRP 83, 272–278 (2019). http://www.sciencedirect.com/science/article/pii/S2212827119306948

  8. Research and Markets: Collaborative Robot (Cobot) Market by Payload, Component (End Effectors, Controllers), Application (Handling, Assembling & Disassembling, Dispensing, Processing), Industry (Electronics, Furniture & Equipment), and Geography – Global Forecast to 2026. https://www.marketsandmarkets.com/Market-Reports/collaborative-robot-market-194541294.html

  9. Malleson, C., Collomosse, J., Hilton, A.: Real-time multi-person motion capture from multi-view video and IMUs. Int. J. Comput. Vis. (2019). https://doi.org/10.1007/s11263-019-01270-5

  10. Manns, M., Fischer, K., Du, H., Slusallek, P., Alexopoulos, K.: A new approach to plan manual assembly. Int. J. Comput. Integr. Manuf., 1–14 (2018). https://www.tandfonline.com/doi/full/10.1080/0951192X.2018.1466396

  11. Manns, M., Mengel, S., Mauer, M.: Experimental effort of data driven human motion simulation in automotive assembly. Procedia CIRP 44, 114–119 (2016)

    Article  Google Scholar 

  12. Manns, M., Otto, M., Mauer, M.: Measuring motion capture data quality for data driven human motion synthesis. Procedia CIRP 41, 945–950 (2016)

    Article  Google Scholar 

  13. Mohammed, O., Bailly, G., Pellier, D.: Acquiring human-robot interaction skills with transfer learning techniques. In: Proceedings of the Companion of the 2017 ACM/IEEE International Conference on Human-Robot Interaction, HRI 2017, pp. 359–360. ACM, New York (2017). https://doi.org/10.1145/3029798.3034823

  14. MOSIM: Github repository codes of MOSIM framework (February 2021). https://github.com/Daimler/MOSIM. Accessed 22 Oct 2020

  15. Reed, K.B., Peshkin, M.A.: Physical collaboration of human-human and human-robot teams. IEEE Trans. Haptics 1(2), 108–120 (2008). Conference Name: IEEE Transactions on Haptics

    Google Scholar 

  16. Roetenberg, D., Luinge, H., Slycke, P.: Xsens MVN: full 6DOF human motion tracking using miniature inertial sensors, p. 7 (2009)

    Google Scholar 

  17. Tirupachuri, Y., Nava, G., Rapetti, L., Latella, C., Darvish, K., Pucci, D.: Recent advances in human-robot collaboration towards joint action. arXiv:2001.00411 [cs] (January 2020). Version: 1

  18. Tuli, T.B., Kohl, L., Chala, S.A., Manns, M., Ansari, F.: Knowledge-Based Digital Twin for Predicting Interactions in Human-Robot Collaboration, Västerås, Sweden (September (2021)

    Google Scholar 

  19. Vesper, C.: How to support action prediction: Evidence from human coordination tasks. In: The 23rd IEEE International Symposium on Robot and Human Interactive Communication, pp. 655–659 (August 2014). ISSN 1944-9437

    Google Scholar 

  20. van Zoelen, E.M., Barakova, E.I., Rauterberg, M.: Adaptive leader-follower behavior in human-robot collaboration. In: 2020 29th IEEE International Conference on Robot and Human Interactive Communication (RO-MAN), pp. 1259–1265 (August 2020). ISSN 1944-9437

    Google Scholar 

Download references

Acknowledgments

The authors would like to acknowledge the financial support by the Federal Ministry of Education and Research of Germany within the ITEA3 project MOSIM (grant number: 01IS18060AH), and by the European Regional Development Fund (EFRE) within the project SMAPS (grant number: 0200545).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Tadele Belay Tuli .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Tuli, T.B., Manns, M., Jonek, M. (2022). Understanding Shared Autonomy of Collaborative Humans Using Motion Capture System for Simulating Team Assembly. In: Andersen, AL., et al. Towards Sustainable Customization: Bridging Smart Products and Manufacturing Systems. CARV MCPC 2021 2021. Lecture Notes in Mechanical Engineering. Springer, Cham. https://doi.org/10.1007/978-3-030-90700-6_59

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-90700-6_59

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-90699-3

  • Online ISBN: 978-3-030-90700-6

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics