Next Article in Journal
Portable Gait Lab: Estimating Over-Ground 3D Ground Reaction Forces Using Only a Pelvis IMU
Next Article in Special Issue
Analysis on the Possibility of Eliminating Interference from Paraseismic Vibration Signals Induced by the Detonation of Explosive Materials
Previous Article in Journal
Low-Complexity Design and Validation of Wireless Motion Sensor Node to Support Physiotherapy
Previous Article in Special Issue
Application of Wireless Accelerometer Mounted on Wheel Rim for Parked Car Monitoring
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Industrial Robot Control by Means of Gestures and Voice Commands in Off-Line and On-Line Mode

Faculty of Mechatronics, Armament and Aerospace, Military University of Technology, Kaliskiego 2 Street, 00-908 Warsaw, Poland
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(21), 6358; https://doi.org/10.3390/s20216358
Submission received: 20 October 2020 / Revised: 3 November 2020 / Accepted: 5 November 2020 / Published: 7 November 2020

Abstract

:
The paper presents the possibility of using the Kinect v2 module to control an industrial robot by means of gestures and voice commands. It describes the elements of creating software for off-line and on-line robot control. The application for the Kinect module was developed in the C# language in the Visual Studio environment, while the industrial robot control program was developed in the RAPID language in the RobotStudio environment. The development of a two-threaded application in the RAPID language allowed separating two independent tasks for the IRB120 robot. The main task of the robot is performed in Thread No. 1 (responsible for movement). Simultaneously, Thread No. 2 ensures continuous communication with the Kinect system and provides information about the gesture and voice commands in real time without any interference in Thread No. 1. The applied solution allows the robot to work in industrial conditions without the negative impact of the communication task on the time of the robot’s work cycles. Thanks to the development of a digital twin of the real robot station, tests of proper application functioning in off-line mode (without using a real robot) were conducted. The obtained results were verified on-line (on the real test station). Tests of the correctness of gesture recognition were carried out, and the robot recognized all programmed gestures. Another test carried out was the recognition and execution of voice commands. A difference in the time of task completion between the actual and virtual station was noticed; the average difference was 0.67 s. The last test carried out was to examine the impact of interference on the recognition of voice commands. With a 10 dB difference between the command and noise, the recognition of voice commands was equal to 91.43%. The developed computer programs have a modular structure, which enables easy adaptation to process requirements.

1. Introduction

The development of electronics, especially sensorics, results in a constant change in the way people interact with electronic devices. An example of using gestures and voice commands to operate devices can today be TVs, computers, building automation systems, mobile robots, etc. The development of electronics and vision systems contributes to the development of a new way of human interaction with devices, which creates completely new possibilities for designing and applying new computer applications. This is particularly evident in interfaces to video surveillance and game applications. A major advantage of using gestures and speech in interaction with the machine is the possibility of dynamic adaptation. The movements of the human body can be considered as segments [1] that express a specific meaning in specific time periods. Hence, during a conversation, people often gesture, emphasizing the meaning of spoken words. Therefore, sequences of gestures can indicate to the machine the next actions to be performed. By adding voice commands to this, it turns out that the method of communication with machines will become very close to the method used in everyday interpersonal relations. When considering this problem, taking into account robotics, it should be noted that the use of gestures and voice commands will make it easier to program and operate very complex devices even for beginners.
Already today, Industry 4.0 integrates people and digitally controlled machines with the Internet and information technologies in various fields of technology [2,3]. On the one hand, the information flow is carried out at the level of the production process between the machines; on the other hand, the process information obtained supports company management systems, which makes it easier to forecast production efficiency.
During the implementation of industrial processes, it is necessary to take into account many technological and hardware parameters [4,5,6,7]. A great challenge is to achieve high productivity and product quality in relation to the costs incurred for the construction of production lines and production itself. Therefore, it is necessary to develop, as much as possible, a flexible production system that would ensure that orders are met today, and in the future, this would allow adapting to changing customer needs. Techniques and tools are sought to analyse the current technological processes with the possibility of anticipating future use of the system [8,9].
Research efforts are being carried out worldwide to highlight the importance of human-machine integration in various aspects. An example is the study [10], which highlighted the importance of human-machine integration in the context of obtaining data for reliability analysis.
In the near future, vision systems used for many years in industrial applications may be used for programming industrial robots or may give the possibility of easier cooperation between the operator and machines at the workstation [11]. Traditional robot programming with the use of dedicated teaching panels slowly becomes a rarity and is replaced by computer programming [12,13,14]. Thus, the question will arise as to what form the programming of robots will take in the future and whether it is possible to create flexible programming that will allow the operator to interact with the robots using natural tools available to man, such as body movement (gestures) or voice. In research laboratories, applications are created that bring us closer to this goal. Developments on programming and controlling robots using gestures and voice are carried out in many fields, e.g., industrial applications (assembly robots), transport (mobile robots), and home use (cleaning robots) [15]. An example can be projects in which images are used to recognize hand gestures in order to control a mobile robot through a specially made control circuit [16]. Of course, the engineer’s or operator’s safety is important in this case, but modern robots today are equipped with safety systems that allow for full collaboration between robots and people [17].
The development and testing of industrial robot control applications can take place in virtual reality in off-line mode. Robot manufacturers offer environments for programming and modelling robots (e.g., RobotStudio by ABB, ROBOGUIDE by FANUC, RT ToolBox by Mitsubishi Electric, KUKA SimPro by KUKA). The modern approach allows creating a virtual controller, which is a digital twin of its real counterpart. This means that it has exactly the same functionality and that the way and principle of its operation almost correspond 100% to the operation of the real device [18]. This gives great possibilities especially because of the functional analysis of software under development. Many researchers use the Kinect sensor for gesture control of robots [19,20,21,22,23,24,25]. An important part of robot programming is preparing an optimal trajectory and gripping points [26,27,28,29,30,31]. An interesting use of the Kinect sensor is its use as a 3D scanning sensor mounted on a FANUC R2000 industrial robot [32]. This approach allows the scanning of large objects on the one hand, but on the other, it is flexible and easily reprogrammable. 3D scanners can also be used to scan larger areas, such as robotic production cells or even entire buildings [33,34]. The limitations of the sensor’s design must be taken into account, as the use of the IR detector requires appropriate environmental conditions to eliminate interference. This solution solves many important problems, but most of it is limited to control by gestures only, without considering the possibility to control voice commands as well. Moreover, it seems to be advisable to implement solutions allowing creating digital stations and using virtual environments.
This paper presents a method of controlling an industrial robot with the use of gestures and voice commands. The solution of the problem required the development of a computer program in the C# language in the Visual Studio environment, which allows obtaining data from the Kinect V2 sensor and is responsible for communication with the application developed for the industrial robot [35]. During the development of a program in Visual Studio, the ABB.Robotics.Controllers library was used to connect directly to the robot controller, set the robot’s operating parameters, and run the robot control program. The Microsoft.Kinect library was used to detect, connect to the application, and receive and interpret data from Kinect V2. This library has the types of variables and mechanisms to operate this device. Voice control of the robot is provided by the System.Speech.Recognition library. The TCP/IP protocol and the System.Net library are responsible for proper communication and smooth robot control. The user interface is based on a WPF (Windows Presentation Foundation) application.
The application for the IRB 120 robot from ABB company was developed in the RAPID programming language. It was decided that the application would be two-threaded. The main thread is responsible for the execution of the robot’s movements and logic related to the production task. The second thread, working in the background of the main thread, is responsible for communication via Ethernet with the application supporting the Kinect sensor.
The prepared software enables testing of the application (e.g., communication of devices and realization of production tasks) and the simulation of the robot’s work in off-line mode with the use of the digital twin and on-line control of the real robot. Tests of the real robot allowed validating the correctness of digital twin’s operation.
Section 2 contains a project of the robotic test station with its digital twin. Section 3 presents a description and the principle of operation of the control programs and the results of the tests carried out. Section 4 summarizes the results of our research.

2. Station Design

Tests were carried out on a research station in the Robotics Laboratory of the Military University of Technology. The station consists of:
  • The ABB six-axis IRB 120 robot with the IRC5 Compact controller (Figure 1),
  • The Kinect V2 sensor (Figure 2),
  • A PC.

2.1. ABB IRB 120 Robot with the IRC5 Compact Controller

The IRB 120 is one of ABB’s smallest industrial robots. It weighs 25 kg and can handle loads up to 3 kg with a range of 580 mm [36].

2.2. Kinect V2 Sensor

The Kinect V2 sensor consists of three components, an RGB camera, a depth sensor (emitter and infrared camera), and a set of four microphones [37], as shown in Figure 2, which give it the functionality to control it with gestures and speech.
A depth sensor is used to track a person’s movement. The infrared emitter sends an infrared beam, which is distorted upon impact with objects or people in its path and is recorded by the infrared camera. From the read data, a room image is created, and the user or users are identified. A set of microphones is used to collect voice information that can be used as control commands.
The Kinect V2 “sees” a person as a set of 25 points (Figure 2b). Their interpretation allows tracking and reading the movements of a person or group of up to six people. Thanks to the points, it is possible to assess in which position a person is currently in front of the device. From these coordinates, it is possible to separate the required coordinates ( X n 1 , X n , Y n 1 , Y n ) , which correspond to the right and left hand. Vector Equations (1)–(3) can be used to convert these coordinates to joint angles [21].
A = X n X n 1
B = Y n Y n 1
θ = A r c T a n ( A , B )
where θ is the angle between two adjacent joints, X n 1 , X n the X coordinates of the joints, and Y n 1 , Y n the Y coordinates of the joints.

2.3. PC

A computer with Windows 10 was used to control an industrial robot with gestures and voice commands. Thanks to the developed control programs, the computer provided support for the Kinect sensor (Visual Studio) and communication with the robot controller (RobotStudio). Kinect for Windows SDK 2.0 (Microsoft.Kinect library) was used to operate the Kinect sensor, while PC SDK 6.07 (ABB.Robotics.Controllers) libraries were used to operate the robot controller.
The key issue for this project was the detection of human body points using the Microsoft Kinect sensor and the use of microphones. A full image processing logic was developed to control the industrial robot using gestures. In addition, a set of microphones was used to control the robot using voice commands.

2.4. Test Station and Its Digital Twin

The flowchart of the station is shown in Figure 3.
One of our aims was to develop the digital twin of the real robotic position (Figure 4). This enabled the creation, modification, and testing of software in a safe way, without the need to work on-line with the real robot.
The backup of the real robot was used to create a digital twin system. In the RobotStudio environment, and the backup generated a robot system that runs on a virtual controller. Such an approach allowed reproducing the real robot controller with almost 100% accuracy. This means that the robot model in a virtual environment has the functionality and features of a real device. This enables the robot model to be programmed taking into account all the capabilities and limitations of the real robot, and the results obtained during the simulation were reliable.

3. Application Design

The integration of devices included in the station (IRB 120 robot, Kinect V2 sensor, computer) required the development of applications enabling the operation of the Kinect sensor and the implementation of an industrial robot control process, as well as ensuring communication between devices (Figure 5).
A control application in the C# language was developed, which was responsible for image processing and communication among the Kinect controller, computer, and robot controller [38,39]. The application was also equipped with a graphical user interface, allowing for free operation without the need to interfere with the program code.
The Microsoft.Kinect library was used to operate the sensor, which allows detecting and connecting to the connected device [40]. For the purposes of the application, an object representing the Kinect was prepared, which made it possible to connect the device with the application being developed. In the next step, an object representing a human being as a whole was declared, and the created logic and software functions allowed to searching for objects representing elements of the human body in the data sent by the Kinect sensor. Next, the authors’ application divided the obtained object into components representing characteristic points of the body, e.g., elbows, shoulders, head, and the state of the hand (open, closed). After recognizing the body of the operator, the position of individual points in relation to each other (Figure 2b) was calculated. The operator’s skeleton is displayed on the user interface as a mirror image to facilitate control, and it was done by developing a sub-program using the points received from the Kinect. The application scaled the received image and displayed it as characteristic points between which lines were drawn, creating a skeleton representing the operator. Based on the operator object received from the Kinect, gestures were proposed to give commands to the robot. Depending on the position of the arms and palms, the developed application allows performing the set of tasks according to the work algorithm.
To implement voice control, the Kinect V2 is connected to a computer and set in Windows 10 as the default microphone (another external microphone can be used). The software was developed using the System.Speech.Recognition library [41]. In the first step, a voice command recognition engine was developed, and the words were declared as text, which were then converted into a dictionary. The prepared dictionary was uploaded to the speech recognition engine, which allows the detection of key words as operator voice commands. On the basis of the recognized commands, a part of the application in which the commands are assigned to the signals controlling the robot was developed.
The application developed in the Visual Studio environment communicates with the robot via Ethernet using the TCP/IP data transmission protocol (Figure 5). The ABB.Robotics.Controllers and System.Net libraries were used to create the application [42]. This application uses ABB libraries to send data to the Robot Communication Run Time module, which is responsible for code synchronization and data transfer in both directions. The authors assumed that the application would allow connecting to all robot controllers in the Ethernet network. In order to make this possible, a search engine was developed for the controllers available in the network. The class implemented in the code allows connecting to the selected controller. The developed application allows direct access to the robot controller before starting the control program, setting the data values and selected parameters of the robot, and uploading them to the controller, e.g., start points of the robot program, manipulator speed, and accuracy. It is also possible to start or stop the robot control program. The next routine allows influencing the robot’s operation also while it is performing its program. This gives the possibility of sending and receiving signals, e.g., sending control signals and receiving the current position of the robot’s manipulator.
The control program for IRB 120 was developed in the RAPID language in the RobotStudio environment. The program is a two-threaded application in which:
  • Thread 1 (Main) is responsible for setting the parameters, interpreting the robot’s direction of motion, and controlling the robot’s movements.
  • Thread 2 is responsible for the TCP/IP communication between the robot and the computer and transferring the data received from the Kinect sensor to the main thread.
This made it possible to make the robot’s motions independent of the actions responsible for data exchange. The threads operate asynchronously based on the same global data (persistent data (PERS)). Thus, if the value of data in one thread changes, it is automatically updated in the other thread. This is of key importance in the control process, because this solution allows controlling an industrial robot using gestures and voice commands without visible delays.

3.1. User Interface

In addition to the full functionality of this application, which allows for controlling the industrial robot by means of gestures and voice commands, an intuitive graphical interface was developed, which allows for easy startup and control of the station.
The user interface was developed in Microsoft Visual Studio 2017 in the C# language. This application enables the user to control an industrial robot by hand movement in 3D space (hand tracking and gestures) (Figure 6) and to control the robot by voice commands.

3.2. Application Tests

The startup tests of the application were performed using the digital twin in the RobotStudio environment (Figure 4a). The developed digital twin allowed for:
  • launching all prepared applications on one computer. The virtual station model was launched in the RobotStudio environment, and communication with the application supporting the Kinect sensor and with the user interface was carried out using the localhost;
  • launching the prepared applications on two computers. The virtual station model was launched in the RobotStudio environment on one computer(simulation of the operation of a real station). The application supporting the Kinect sensor and the user interface was launched on the second computer. The computers were connected to the local Ethernet.
During the tests, the correctness of the robot’s task performance was verified both on the basis of performed gestures and voice commands. The adopted solution allowed the execution of the tests. Validation of the application’s correctness was carried out on the real station (Figure 4b and Figure 7).

4. Results and Discussion

A number of tests were carried out to verify the correctness of the robot’s reaction to gestures and voice commands. Tests were conducted using a digital twin and a real robot. Both single movements and complex tasks (e.g., pick and place) were tested.
In the case of robot control by means of gestures, tests were carried out on the performance of commands by the robot in the form of motion in selected directions (Table 1), following the operator’s hand (Figure 7), performing a complex task, and controlling the robot’s input signals. No communication errors or unrecognised commands were observed. The dynamics of the Kinect sensor and, above all, the acceleration and speed capability of the robot are limited when the robot follows the hand movement. This type of control seems to be useful when tasks need to be carried out remotely (in a place that is difficult to reach or far away), and the time of task completion does not play a significant role here. Gestures seems to be a more promising control method than hand tracking, because it may execute complex tasks for the robot (e.g., assembly, sorting). The main advantage in this case is that the operator can start the robot, indicate the task, or turn off the robot without having to use an additional control panel.
Tests of the delay of the robot’s reaction to voice commands were conducted. Selected test results are presented below (Table 2 and Table 3).
Figure 8 shows the delay of the robot’s reaction to the voice commands on the real station.
Figure 9 shows the delays expressed in seconds for voice control commands on the virtual station.
Table 2 and Table 3 indicate a slower robot response in the RobotStudio environment. There is a difference of 0.92 s between the fastest average response of the real robot as compared to a virtual robot. In both cases, the robot best responds to commands to control along the y-axis. The average response rate for the actual device is 1.74 s and for the virtual one 2.41 s. This may be due to the fact that Robot Studio is not working in real time and the insufficient computing capacity of the computer used was affecting the processing time. Delay may also be caused by the fact that communication was performed using a TCP/IP over a network and that the robot controller may have tasks of higher priority to perform first.
The next study evaluates the influence of sound interference intensity (20 dB, 40 dB, 60 dB) on the ability to control the robot by voice commands. The “Sound Master” application was used to measure the interference. The measuring device was located next to the Kinect sensor. A music track without vocals was used as a disturbance. The commands for the robot were given in a calm voice of 70 dB in intensity. A number of tests were carried out on a virtual and real station, changing the robot’s parameters. For 20 dB and 40 dB interference, there were no problems with the performance of commands. The tables below show the results for 60 dB motion commands.
The results presented in Table 4 show that commands of 70 dB in intensity can be incomprehensible for a robot system equipped with the Kinect sensor if the interference is at 60 dB. The achieved recognition rate of voice commands is equal to 91.43%. This value might be enough for performing laboratory tests, but is definitely not acceptable in industrial environments. However, it should be noted that in the case of simple commands (for the virtual and real system), the “plus y” command caused the problem, while other commands were recognized correctly. Therefore, it is important that the selected commands are easily and clearly identified by the system, especially in industrial conditions (with noise).

5. Conclusions

Human interaction with an industrial robot using a vision system equipped with microphones seems to be an interesting and promising solution for future robot programming. Teaching the trajectory of industrial robots with gestures and voice commands may prove to be one of the methods of robot programming in the future. The combination of gestures and voice commands also seems to be an interesting issue in terms of robot control, especially since the advantage of using gestures and voice commands in interaction with the machine is the ability to adapt dynamically. For humans, this form of communication seems to be the most intuitive, as we learn from childhood to use gestures for more accurate expression of our feelings, emotions, and needs. In our lives (in direct contacts), speech and gestures coexist, and therefore, they should also be combined in interactions with machines. As recent trends indicate, in the near future, this type of control will become a standard in cars, where the leading manufacturers already offer such functionality. An example is the handling of navigation and multimedia systems. Control of this type has many supporters, although there are also opinions that voice and gesture control systems do not work precisely and unambiguously enough. This is mainly due to the appearance of interferences (e.g., noise or sounds of music in a moving car) and inefficient algorithms of voice and gesture analysis and processing. Currently developed algorithms are getting better and better, which allows us to think that they will also be used in factories. Expanding the systems of industrial robots and their control programs with the possibility to control by means of gestures and voice commands will make it easier for operators of production stations to communicate with machines and will make them able to operate extended stations after a short training. It should be expected that in the future, this type of control will especially find its application in areas where humans will cooperate with a robot (e.g., collaborating robots (cobots)). The possibility of easier communication with cobots (without additional distractions for operators) will allow for closer and faster cooperation, similar to the level of cooperation between cooperating people.
This paper focuses on the presentation of prepared software, which allows controlling an industrial robot by means of gestures and voice commands. ABB IRB 120 robot was used as an example, but the developed software can be easily integrated with other robots. For the IRB 120 robot, we developed a multithreaded control program that allows for a flexible tasks realization. Our aim was to develop an inexpensive and universal solution. Tests showed that gesture control allows for closer cooperation between man and machine. Additional voice commands extend the possibilities of controlling the robot without the need to focus all the operator’s attention on it. The presented test results prove that the recognition of voice commands depends on the noise level. It was noted that the 10 dB difference between command and noise provided 91.43% recognition of the commands.
The paper also presents the possibility of performing tests based on digital twins. Already today, such an approach allows creating control programs for robots and conducting analyses of the work of robotic stations, even entire production lines and factories. The simulation results are particularly important for integrators (at the start of production) and managers (who manage production and conduct analyses of market changes). The research carried out, using a real robotic station and a virtual robotic system, shows that in the case of voice handling, the average difference between real and virtual robot performance was 0.67 s. This difference may result from the fact that part of the API executes on a non-real-time operating system, and the communication is done using a TCP/IP over a network. Furthermore, the robot controller sometimes has tasks of higher priority to perform first. Such analyses allow for a quick response to market changes and the introduction of changes in production. The use of digital twins allows preparing and testing the introduced changes without stopping production.
In the next stage of research, the industrial robot will be controlled by means of gestures and voice commands taking into account the robot’s safety zones. Moreover, with the use of digital twins, tests of controlling the robot by means of gestures and voice commands in virtual reality with HTC VIVE Cosmos goggles will be conducted.

Author Contributions

Conceptualization, W.K. and P.B.; methodology, W.K. and P.B.; software, W.K. and P.B.; validation, P.B., S.B., and J.P.; formal analysis, W.K. and J.P.; investigation, P.B. and W.K.; resources, S.B., W.K., and J.P.; data curation, W.K., P.B., and J.P.; writing, original draft preparation, S.B., W.K., and P.B.; writing, review and editing, S.B., W.K., and J.P.; visualization, W.K. and S.B.; supervision, W.K. and J.P.; project administration, W.K. All authors read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Peursum, P.; Bui, H.; Venkatesh, S.; West, G. Human action segmentation via controlled use of missing data in HMMs. In Proceedings of the 17th International Conference on Pattern Recognition, Cambridge, UK, 26–26 August 2004; Volume 4, pp. 440–445. [Google Scholar] [CrossRef] [Green Version]
  2. Kaczmarek, W.; Panasiuk, J.; Borys, S.; Pobudkowska, A.; Majsterek, M. Analysis of the kinetics of swimming pool water reaction in analytical device reproducing its circulation on a small scale. Sensors 2020, 20, 17. [Google Scholar] [CrossRef] [PubMed]
  3. Wortmann, A.; Barais, O.; Combemale, B.; Wimmer, M. Modeling languages in Industry 4.0: An extended systematic mapping study. Softw. Syst. Model. 2020, 19, 67–94. [Google Scholar] [CrossRef] [Green Version]
  4. Bect, P.; Simeu-Abazi, Z.; Maisonneuve, P.L. Identification of abnormal events by data monitoring: Application to complex systems. Comput. Ind. 2015, 68, 78–88. [Google Scholar] [CrossRef]
  5. Konieczny, Ł; Burdzik, R. Comparison of Characteristics of the Components Used in Mechanical and Non-Conventional Automotive Suspensions. Solid State Phenom. 2013, 210, 26–31. [Google Scholar] [CrossRef]
  6. Gasiorek, D.; Baranowski, P.; Malachowski, J.; Mazurkiewicz, L.; Wiercigroch, M. Modelling of guillotine cutting of multi-layered aluminum sheets. J. Manuf. Process. 2018, 34, 374–388. [Google Scholar] [CrossRef]
  7. Borys, S.; Kaczmarek, W.; Laskowski, D. Selection and Optimization of the Parameters of the Robotized Packaging Process of One Type of Product. Sensors 2020, 20, 5378. [Google Scholar] [CrossRef]
  8. Sekala, A.; Gwiazda, A.; Kost, G.; Banas, W. Modelling of a production system using the multi-agent network approach. IOP Conf. Ser.: Mater. Sci. Eng. 2018, 400, 052009. [Google Scholar] [CrossRef]
  9. Vera-Baquero, A.; Colomo-Palacios, R.; Molloy, O. Towards a Process to Guide Big Data Based Decision Support Systems for Business Processes. Procedia Technol. 2014, 16, 11–21. [Google Scholar] [CrossRef] [Green Version]
  10. Esmaeilian, B.; Behdad, S.; Wang, B. The evolution and future of manufacturing: A review. J. Manuf. Syst. 2016, 39, 79–100. [Google Scholar] [CrossRef]
  11. Prusaczyk, P.; Kaczmarek, W.; Panasiuk, J.; Besseghieur, K. Integration of robotic arm and vision system with processing software using TCP/IP protocol in industrial sorting application. In Proceedings of the 15th Conference on Computational Technologies in Engineering, Svratka, Czech Republic, 13–16 May 2019; p. 020032. [Google Scholar] [CrossRef]
  12. Kaczmarek, W.; Panasiuk, J.; Borys, S. Środowiska Programowania Robotów; PWN Scientific Publishing House: Warsaw, Poland, 2017. [Google Scholar]
  13. Kaczmarek, W.; Panasiuk, J. Robotyzacja Procesow Produkcyjnych; PWN Scientific Publishing House: Warsaw, Poland, 2017. [Google Scholar]
  14. Kaczmarek, W.; Panasiuk, J. Programowanie Robotów Przemysłowych; PWN Scientific Publishing House: Warsaw, Poland, 2017. [Google Scholar]
  15. Park, C.B.; Lee, S.W. Real-time 3D pointing gesture recognition for mobile robots with cascade HMM and particle filter. Image Vis. Comput. 2011, 29, 51–63. [Google Scholar] [CrossRef]
  16. Faudzi, A.M.; Ali, M.H.K.; Azman, M.A.; Ismail, Z.H. Real-time Hand Gestures System for Mobile Robots Control. Procedia Eng. 2012, 41, 798–804. [Google Scholar] [CrossRef] [Green Version]
  17. Rosenstrauch, M.J.; Pannen, T.J.; Krüger, J. Human robot collaboration—Using Kinect v2 for ISO/TS 15066 speed and separation monitoring. Procedia CIRP 2018, 76, 183–186. [Google Scholar] [CrossRef]
  18. Rosen, R.; von Wichert, G.; Lo, G.; Bettenhausen, K.D. About The Importance of Autonomy and Digital Twins for the Future of Manufacturing. IFAC-PapersOnLine 2015, 48, 567–572. [Google Scholar] [CrossRef]
  19. Bouteraa, Y.; Ben Abdallah, I. A gesture-based telemanipulation control for a robotic arm with biofeedback-based grasp. Ind. Robot Int. J. 2017, 44, 575–587. [Google Scholar] [CrossRef]
  20. Bernier, E.; Chellali, R.; Thouvenin, I.M. Human gesture segmentation based on change point model for efficient gesture interface. In Proceedings of the 2013 IEEE RO-MAN, Gyeongju, Korea, 26–29 August 2013; pp. 258–263. [Google Scholar] [CrossRef]
  21. Ghonge, E.P.; Kulkarni, M.N. Gesture based control of IRB1520ID using Microsoft’s Kinect. In Proceedings of the 2017 IEEE 2nd International Conference on Communication and Electronics Systems (ICCES), Coimbatore, India, 19–20 October 2017; pp. 355–358. [Google Scholar] [CrossRef]
  22. Jingbiao, L.; Huan, X.; Zhu, L.; Qinghua, S. Dynamic gesture recognition algorithm in human computer interaction. In Proceedings of the 2015 IEEE 16th International Conference on Communication Technology (ICCT), Hangzhou, China, 18–20 October 2015; pp. 425–428. [Google Scholar] [CrossRef]
  23. Popov, V.; Ahmed, S.; Shakev, N.; Topalov, A. Gesture-based Interface for Real-time Control of a Mitsubishi SCARA Robot Manipulator. IFAC-PapersOnLine 2019, 52, 180–185. [Google Scholar] [CrossRef]
  24. Torres, S.H.M.; Kern, M.J. 7 DOF industrial robot controlled by hand gestures using microsoft Kinect v2. In Proceedings of the 2017 IEEE 3rd Colombian Conference on Automatic Control (CCAC), Cartagena, Colombia, 18–20 October 2017; pp. 1–6. [Google Scholar] [CrossRef]
  25. Moe, S.; Schjolberg, I. Real-time hand guiding of industrial manipulator in 5 DOF using Microsoft Kinect and accelerometer. In Proceedings of the 2013 IEEE RO-MAN, Gyeongju, Korea, 26–29 August 2013; pp. 644–649. [Google Scholar] [CrossRef]
  26. Jha, A.; Chiddarwar, S.S.; Alakshendra, V.; Andulkar, M.V. Kinematics-based approach for robot programming via human arm motion. J. Braz. Soc. Mech. Sci. Eng. 2017, 39, 2659–2675. [Google Scholar] [CrossRef]
  27. Duque, D.A.; Prieto, F.A.; Hoyos, J.G. Trajectory generation for robotic assembly operations using learning by demonstration. Robot. Comput. Integr. Manuf. 2019, 57, 292–302. [Google Scholar] [CrossRef]
  28. Jha, A.; Chiddarwar, S.S.; Andulkar, M.V. An integrated approach for robot training using Kinect and human arm kinematics. In Proceedings of the 2015 IEEE International Conference on Advances in Computing, Communications and Informatics (ICACCI), Kochi, India, 10–13 August 2015; pp. 216–221. [Google Scholar] [CrossRef]
  29. Chen, C.S.; Chen, P.C.; Hsu, C.M. Three-Dimensional Object Recognition and Registration for Robotic Grasping Systems Using a Modified Viewpoint Feature Histogram. Sensors 2016, 16, 1969. [Google Scholar] [CrossRef] [Green Version]
  30. Zhou, Y.; Chen, M.; Du, G.; Zhang, P.; Liu, X. Intelligent grasping with natural human-robot interaction. Ind. Robot Int. J. 2018, 45, 44–53. [Google Scholar] [CrossRef]
  31. Wang, F.; Liang, C.; Ru, C.; Cheng, H. An Improved Point Cloud Descriptor for Vision Based Robotic Grasping System. Sensors 2019, 19, 2225. [Google Scholar] [CrossRef] [Green Version]
  32. Panasiuk, J.; Kaczmarek, W.; Besseghieur, K.; Borys, S.; Siwek, M.; Goławski, K. Study of the 3D scanning accuracy using industrial robot. In Proceedings of the 24th International Conference—Engineering Mechanics, Svratka, Czech Republic, 14–17 May 2018; pp. 641–644. [Google Scholar] [CrossRef] [Green Version]
  33. Dybedal, J.; Aalerud, A.; Hovland, G. Embedded Processing and Compression of 3D Sensor Data for Large Scale Industrial Environments. Sensors 2019, 19, 636. [Google Scholar] [CrossRef] [Green Version]
  34. Ullo, S.L.; Zarro, C.; Wojtowicz, K.; Meoli, G.; Focareta, M. LiDAR-Based System and Optical VHR Data for Building Detection and Mapping. Sensors 2020, 20, 1285. [Google Scholar] [CrossRef] [Green Version]
  35. Perkins, B.; Hammer, J.V.; Reid, J.D. Beginning C# 7 Programming with Visual Studio® 2017; Wiley: New York, NY, USA, 2018. [Google Scholar] [CrossRef]
  36. Product Specification IRB 120; ABB AB Robotics: Västerås, Sweden. Available online: https://library.e.abb.com/public/7139d7f4f2cb4d0da9b7fac6541e91d1/3HAC035960%20PS%20IRB%20120-en.pdf (accessed on 15 August 2020).
  37. KINECT for Windows. Human Interface Guidelines v2.0. Microsoft Corporation, USA. Available online: https://download.microsoft.com/download/6/7/6/676611B4-1982-47A4-A42E-4CF84E1095A8/KinectHIG.2.0.pdf (accessed on 15 August 2020).
  38. Albahari, J.; Albahari, B. C# 7.0 in a Nutshell: The Definitive Reference; O’Reilly Media Inc.: Sevastopol, CA, USA, 2017. [Google Scholar]
  39. Troelsen, A.; Japikse, P. C# 6.0 and the .NET 4.6 Framework; Apress: Berkeley, CA, USA, 2015. [Google Scholar] [CrossRef]
  40. Webb, J.; Ashley, J. Beginning Kinect Programming with the Microsoft Kinect SDK; Apress: Berkeley, CA, USA, 2012. [Google Scholar] [CrossRef]
  41. Greenberg, S.; Ainsworth, W.; Popper, A.; Fay, R. Speech Processing in the Auditory System; Springer Handbook of Auditory Research; Springer: New York, NY, USA, 2004; Volume 18. [Google Scholar] [CrossRef]
  42. PC SDK API Reference. ABB AB Robotics: Västerås, Sweden. Available online: https://developercenter.robotstudio.com/api/pcsdk/api/ (accessed on 15 August 2020).
Figure 1. General view: (a) ABB IRB 120 robot with FlexPendant; (b) IRC5 Compact robot controller.
Figure 1. General view: (a) ABB IRB 120 robot with FlexPendant; (b) IRC5 Compact robot controller.
Sensors 20 06358 g001
Figure 2. Kinect V2: (a) general view; (b) recognizable body points.
Figure 2. Kinect V2: (a) general view; (b) recognizable body points.
Sensors 20 06358 g002
Figure 3. Flowchart of the station.
Figure 3. Flowchart of the station.
Sensors 20 06358 g003
Figure 4. General view of the robotised station: (a) virtual; (b) real.
Figure 4. General view of the robotised station: (a) virtual; (b) real.
Sensors 20 06358 g004
Figure 5. General diagram of communication between the application and the robot controller.
Figure 5. General diagram of communication between the application and the robot controller.
Sensors 20 06358 g005
Figure 6. Graphical interface of the developed application: 1—help button, 2—program information button, 3—control method list, 4—current controlled axis, 5—current x, y, z axis position field, 6—list of available robots, 7—robot accuracy window, 8—robot speed selection window, 9—start button, 10—operator’s “skeleton” field.
Figure 6. Graphical interface of the developed application: 1—help button, 2—program information button, 3—control method list, 4—current controlled axis, 5—current x, y, z axis position field, 6—list of available robots, 7—robot accuracy window, 8—robot speed selection window, 9—start button, 10—operator’s “skeleton” field.
Sensors 20 06358 g006
Figure 7. Views of the test station: 1—ABB IRB 120 robot, 2—operator, 3—operator interface window.
Figure 7. Views of the test station: 1—ABB IRB 120 robot, 2—operator, 3—operator interface window.
Sensors 20 06358 g007
Figure 8. Measurement of the delay of the robot’s reaction to the voice command: real station.
Figure 8. Measurement of the delay of the robot’s reaction to the voice command: real station.
Sensors 20 06358 g008
Figure 9. Measurement of the delay of the robot’s reaction to the voice command: virtual station.
Figure 9. Measurement of the delay of the robot’s reaction to the voice command: virtual station.
Sensors 20 06358 g009
Table 1. Gestures used in robot control.
Table 1. Gestures used in robot control.
Gesture *Robot ActionOperator’s Body Position
Sensors 20 06358 i001Axis X move forwardRight hand on the left, left hand below shoulder
Sensors 20 06358 i002Axis X move backwardsRight hand on the right, left hand below shoulder
Sensors 20 06358 i003Axis Y move forwardRight hand on the left, left hand between shoulder and head
Sensors 20 06358 i004Axis Y move backwardsRight hand on the right, left hand between shoulder and head
Sensors 20 06358 i005Axis Z move forwardRight hand on the left, left hand above head
Sensors 20 06358 i006Axis Z move backwardsRight hand on the right, left hand above head
Sensors 20 06358 i007Confirm movementClose left hand
* The skeleton field is mirrored.
Table 2. Measurement of the delay of the robot’s reaction to the voice command: real station.
Table 2. Measurement of the delay of the robot’s reaction to the voice command: real station.
Sample NbCommands
“plus x”“minus x”“plus y”“minus y”“plus z”“minus z”stop
Sample 1 (s)1.851.671.551.611.741.681.70
Sample 2 (s)1.931.941.501.472.001.851.60
Sample 3 (s)2.061.781.761.651.801.841.85
Sample 4 (s)2.071.601.691.801.891.791.59
Sample 5 (s)1.671.781.501.521.891.741.75
Average delay (s)1.921.751.601.611.861.781.70
Table 3. Measurement of the delay of the robot’s reaction to the voice command: virtual station.
Table 3. Measurement of the delay of the robot’s reaction to the voice command: virtual station.
Sample NbCommands
“plus x”“minus x”“plus y”“minus y”“plus z”“minus z”“stop”
Sample 1 (s)2.682.452.002.432.482.372.40
Sample 2 (s)2.282.202.492.472.312.402.42
Sample 3 (s)2.462.542.502.002.622.672.50
Sample 4 (s)2.172.492.192.492.552.512.19
Sample 5 (s)2.342.672.172.502.662.402.38
Average delay (s)2.392.472.272.382.522.472.39
Table 4. Effects of 60 dB interference on the robot’s operation.
Table 4. Effects of 60 dB interference on the robot’s operation.
Sample NbCommands
“plus x”“minus x”“plus y”“minus y”“plus z”“minus z”“stop”
Sample 1okokokokokokok
Sample 2okokokokokokok
Sample 3okokokokokokok
Sample 4okokxokokokok
Sample 5okokxokokokok
Sample 6okokxokokokok
Sample 7okokxokokokok
Sample 8okokokokxokok
Sample 9okokokokokokok
Sample 10okokxokokokok
Recognition rate (%)1001005010090100100
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Kaczmarek, W.; Panasiuk, J.; Borys, S.; Banach, P. Industrial Robot Control by Means of Gestures and Voice Commands in Off-Line and On-Line Mode. Sensors 2020, 20, 6358. https://doi.org/10.3390/s20216358

AMA Style

Kaczmarek W, Panasiuk J, Borys S, Banach P. Industrial Robot Control by Means of Gestures and Voice Commands in Off-Line and On-Line Mode. Sensors. 2020; 20(21):6358. https://doi.org/10.3390/s20216358

Chicago/Turabian Style

Kaczmarek, Wojciech, Jarosław Panasiuk, Szymon Borys, and Patryk Banach. 2020. "Industrial Robot Control by Means of Gestures and Voice Commands in Off-Line and On-Line Mode" Sensors 20, no. 21: 6358. https://doi.org/10.3390/s20216358

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop