Abstract
Long-distance, high latency teleoperation tasks are difficult, highly stressful for teleoperators, and prone to over-corrections, which can lead to loss of control. At higher latencies, or when teleoperating at higher vehicle speed, the situation becomes progressively worse. To explore potential solutions, this research work investigates two 2D visual feedback-based assistive interfaces (sliding-only and sliding-and-zooming windows) that apply simple but effective video transformations to enhance teleoperation. A teleoperation simulator that can replicate teleoperation scenarios affected by high and adjustable latency has been developed to explore the effectiveness of the proposed assistive interfaces. Three image comparison metrics have been used to fine-tune and optimise the proposed interfaces. An operator survey was conducted to evaluate and compare performance with and without the assistance. The survey has shown that a 900ms latency increases task completion time by up to 205% for an on-road and 147% for an off-road driving track. Further, the overcorrection-induced oscillations increase by up to 718% with this level of latency. The survey has shown the sliding-only video transformation reduces the task completion time by up to 25.53%, and the sliding-and-zooming transformation reduces the task completion time by up to 21.82%. The sliding-only interface reduces the oscillation count by up to 66.28%, and the sliding-and-zooming interface reduces it by up to 75.58%. The qualitative feedback from the participants also shows that both types of assistive interfaces offer better visual situational awareness, comfort, and controllability, and significantly reduce the impact of latency and intermittency on the teleoperation task.
Article PDF
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Data availability
If requested, the authors will share the data collected and used in this paper.
Code availability
If requested, the authors will share the code and Simulink model prepared and used in this paper.
References
Moniruzzaman, M., Rassau, A., Chai, D., Islam, S.M.S.: Robotic teleoperation methods and enhancement techniques: A comprehensive survey. Robotics and Autonomous Systems 150, 103973 (2022). https://doi.org/10.1016/j.robot.2021.103973
Little, C.L., Perry, E.E., Fefer, J.P., Brownlee, M.T., Sharp, R.L.: An interdisciplinary review of camera image collection and analysis techniques, with considerations for environmental conservation social science. Data 5(2), 51 (2020). https://doi.org/10.3390/data5020051
Fong, T., Thorpe, C., Baur, C.: Advanced interfaces for vehicle teleoperation: Collaborative control, sensor fusion displays, and remote driving tools. Auton. Robots 11(1), 77–85 (2001). https://doi.org/10.1023/A:1011212313630
Sheridan, T.B.: Telerobotics. Automatica 25(4), 487–507 (1989). https://doi.org/10.1016/0005-1098(89)90093-9
Choi, J.J., Kim, Y., Kwak, S.S.: The autonomy levels and the human intervention levels of robots: The impact of robot types in human-robot interaction. In: The 23rd IEEE International Symposium on Robot and Human Interactive Communication, pp. 1069–1074. IEEE, Edinburgh (2014). https://doi.org/10.1109/ROMAN.2014.6926394
Dorais, G., Bonasso, R.P., Kortenkamp, D., Pell, B., Schreckenghost, D.: Adjustable autonomy for human-centered autonomous systems. In: Proceedings of the Sixteenth International Joint Conference on Artificial Intelligence Workshop on Adjustable Autonomy Systems, pp. 16–35 (1999)
Fong, T., Thorpe, C., Baur, C.: Robot as partner: Vehicle teleoperation with collaborative control. In: Multi-Robot Systems: From Swarms to Intelligent Automata, pp. 195–202. Springer, Washington DC, USA (2002). https://doi.org/10.1007/978-94-017-2376-3_21
Bruemmer, D.J., Marble, J.L., Dudenhoeffer, D.D., Anderson, M., McKay, M.D.: Mixed-initiative control for remote characterization of hazardous environments. In: Proceedings of the 36th Annual Hawaii International Conference on System Sciences (HICSS), pp. 9–17. IEEE, Big Island, USA (2003). https://doi.org/10.1109/HICSS.2003.1174289
Sellner, B., Heger, F.W., Hiatt, L.M., Simmons, R., Singh, S.: Coordinated multiagent teams and sliding autonomy for large-scale assembly. Proc. IEEE 94(7), 1425–1444 (2006). https://doi.org/10.1109/JPROC.2006.876966
MacKenzie, I.S., Ware, C.: Lag as a determinant of human performance in interactive systems. In: Proceedings of the INTERACT’93 and CHI’93 Conference on Human Factors in Computing Systems, Amsterdam, Netherlands, pp. 488–493 (1993). https://doi.org/10.1145/169059.169431
Chen, J.Y., Haas, E.C., Barnes, M.J.: Human performance issues and user interface design for teleoperated robots. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews) 37(6), 1231–1245 (2007). https://doi.org/10.1109/TSMCC.2007.905819
Matheson, A., Donmez, B., Rehmatullah, F., Jasiobedzki, P., Ng, H.-K., Panwar, V., Li, M.: The effects of predictive displays on performance in driving tasks with multi-second latency: Aiding tele-operation of lunar rovers. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 57, pp. 21–25. SAGE Publications, Los Angeles, CA (2013). https://doi.org/10.1177/1541931213571007
Lane, J.C., Carignan, C.R., Sullivan, B.R., Akin, D.L., Hunt, T., Cohen, R.: Effects of time delay on telerobotic control of neutral buoyancy vehicles. In: Proceedings 2002 IEEE International Conference on Robotics and Automation, vol. 3, pp. 2874–2879. IEEE, Washington DC, USA (2002). https://doi.org/10.1109/ROBOT.2002.1013668
Ellis, S.R., Mania, K., Adelstein, B.D., Hill, M.I.: Generalizeability of latency detection in a variety of virtual environments. In: Proceedings of the Human Factors and Ergonomics Society Annual Meeting, vol. 48, pp. 2632–2636. SAGE Publications, Los Angeles, CA (2004). https://doi.org/10.1177/154193120404802306
Held, R., Efstathiou, A., Greene, M.: Adaptation to displaced and delayed visual feedback from the hand. J. Exp. Psychol. 72(6), 887 (1966). https://doi.org/10.1037/h0023868
Barnes, M., Cosenzo, K., Mitchell, D., Chen, J.: Human robot teams as soldier augmentation in future battlefields: An overview. In: Proceedings of the 11th International Conference of Hum (2005)
Sanzharov, V., Frolov, V., Galaktionov, V.: Survey of nvidia rtx technology. Program. Comput. Softw. 46(4), 297–304 (2020). https://doi.org/10.1134/S0361768820030068
Mahmud, S., Lin, X., Kim, J.-H.: Interface for human machine interaction for assistant devices: a review. In: 10th Annual Computing and Communication Workshop and Conference (CCWC), pp. 0768–0773. IEEE, Las Vegas, NV, USA (2020). https://doi.org/10.1109/CCWC47524.2020.9031244
Khan, A., Sohail, A., Zahoora, U., Qureshi, A.S.: A survey of the recent architectures of deep convolutional neural networks. Artif. Intell. Rev. 53(8), 5455–5516 (2020). https://doi.org/10.1007/s10462-020-09825-6
Al-Garadi, M.A., Mohamed, A., Al-Ali, A.K., Du, X., Ali, I., Guizani, M.: A survey of machine and deep learning methods for internet of things (iot) security. IEEE Commun. Surv. Tutor. 22(3), 1646–1685 (2020). https://doi.org/10.1109/COMST.2020.2988293
Draper, J.V., Kaber, D.B., Usher, J.M.: Telepresence. Hum. Factors 40(3), 354–375 (1998). https://doi.org/10.1518/001872098779591386
Fong, T., Thorpe, C.: Vehicle teleoperation interfaces. Auton. Robots 11(1), 9–18 (2001). https://doi.org/10.1023/A:1011295826834
Fong, T., Thorpe, C., Baur, C.: Advanced interfaces for vehicle teleoperation: Collaborative control, sensor fusion displays, and remote driving tools. Auton. Robots 11(1), 77–85 (2001). https://doi.org/10.1023/A:1011212313630
Nof, S.Y.: Springer Handbook of Automation. Springer, Heidelberg, Germany (2009)
Dybvik, H., Løland, M., Gerstenberg, A., Slåttsveen, K.B., Steinert, M.: A low-cost predictive display for teleoperation: Investigating effects on human performance and workload. Int. J. Hum. Comput. Stud. 145, 102536 (2021). https://doi.org/10.1016/j.ijhcs.2020.102536
Wilde, M., Chan, M., Kish, B.: Predictive human-machine interface for teleoperation of air and space vehicles over time delay. In: 2020 IEEE Aerospace Conference, pp. 1–14. IEEE, Big Sky, MT, USA (2020). https://doi.org/10.1109/AERO47225.2020.9172297
Ha, C., Yoon, J., Kim, C., Lee, Y., Kwon, S., Lee, D.: Teleoperation of a platoon of distributed wheeled mobile robots with predictive display. Auton. Robots 42(8), 1819–1836 (2018)
Wang, C., Zuo, Z., Lin, Z., Ding, Z.: A truncated prediction approach to consensus control of lipschitz nonlinear multiagent systems with input delay. IEEE Trans. Control. Netw. Syst. 4(4), 716–724 (2016). https://doi.org/10.1109/TCNS.2016.2545860
Richard, J.-P.: Time-delay systems: an overview of some recent advances and open problems. Automatica 39(10), 1667–1694 (2003). https://doi.org/10.1016/S0005-1098(03)00167-5
Manitius, A., Olbrot, A.: Finite spectrum assignment problem for systems with delays. IEEE Trans. Autom. Control 24(4), 541–552 (1979). https://doi.org/10.1109/TAC.1979.1102124
Artstein, Z.: Linear systems with delayed controls: A reduction. IEEE Trans. Autom. Control 27(4), 869–879 (1982). https://doi.org/10.1109/TAC.1982.1103023
Bejczy, A.K., Kim, W.S.: Predictive displays and shared compliance control for time-delayed telemanipulation. In: IEEE International Workshop on Intelligent Robots and Systems, Towards a New Frontier of Applications, pp. 407–412. IEEE, Ibaraki, Japan (1990). https://doi.org/10.1109/IROS.1990.262418
Bejczy, A.K., Kim, W.S., Venema, S.C.: The phantom robot: predictive displays for teleoperation with time delay. In: Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Cincinnati, OH, USA, pp. 546–551 (1990). https://doi.org/10.1109/ROBOT.1990.126037
Buzan, F.T., Sheridan, T.B.: A model-based predictive operator aid for telemanipulators with time delay. In: Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, Cambridge, MA, USA, pp. 138–143 (1989). https://doi.org/10.1109/ICSMC.1989.71268
Hirzinger, G., Heindl, J., Landzettel, K.: Predictive and knowledge-based telerobotic control concepts. In: IEEE International Conference on Robotics and Automation, Scottsdale, AZ, USA, pp. 1768–1769 (1989). https://doi.org/10.1109/ROBOT.1989.100231
Witus, G., Hunt, S., Janicki, P.: Methods for ugv teleoperation with high latency communications. In: Unmanned Systems Technology XIII, vol. 8045, p. 80450. SPIE, Orlando, Florida, United States (2011). https://doi.org/10.1117/12.886058
Zheng, Y., Brudnak, M.J., Jayakumar, P., Stein, J.L., Ersal, T.: An experimental evaluation of a model-free predictor framework in teleoperated vehicles. IFAC-PapersOnLine 49(10), 157–164 (2016). https://doi.org/10.1016/j.ifacol.2016.07.513
Zhang, Y., Li, H.: Handling qualities evaluation of predictive display model for rendezvous and docking in lunar orbit with large time delay. In: 2016 IEEE Chinese Guidance, Navigation and Control Conference (CGNCC), pp. 742–747. IEEE, Nanjing, China (2016). https://doi.org/10.1109/CGNCC.2016.7828878
Johnson, D.H.: Signal-to-noise ratio. Scholarpedia 1(12), 2088 (2006). https://doi.org/10.4249/scholarpedia.2088
Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600–612 (2004). https://doi.org/10.1109/TIP.2003.819861
Eikvil, L.: Optical Character Recognition. Citeseer, Princeton, New Jersey, USA (1993)
Brunet, D., Vrscay, E.R., Wang, Z.: On the mathematical properties of the structural similarity index. IEEE Trans. Image Process. 21(4), 1488–1499 (2011). https://doi.org/10.1109/TIP.2011.2173206
Wang, Z., Simoncelli, E.P., Bovik, A.C.: Multiscale structural similarity for image quality assessment. In: The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, vol. 2, pp. 1398–1402. IEEE, Pacific Grove, CA, USA (2003). https://doi.org/10.1109/ACSSC.2003.1292216
Sobel, I., Duda, R., Hart, P., Wiley, J.: Sobel-Feldman Operator. Preprint at https://www.researchgate.net/profile/Irwin-Sobel/publication/285159837. Accessed 20 Apr 2022
Rouse, D.M., Hemami, S.S.: Analyzing the role of visual structure in the recognition of natural image content with multi-scale ssim. In: Human Vision and Electronic Imaging XIII, vol. 6806, pp. 680615. SPIE, San Jose, California, United States (2008). https://doi.org/10.1117/12.768060
Søgaard, J., Krasula, L., Shahid, M., Temel, D., Brunnström, K., Razaak, M.: Applicability of existing objective metrics of perceptual quality for adaptive video streaming. Electronic Imaging 2016(13), 1–7 (2016)
Dosselmann, R., Yang, X.D.: A comprehensive assessment of the structural similarity index. Signal Image Video Process. 5(1), 81–91 (2011). https://doi.org/10.1007/s11760-009-0144-1
Lu, S., Zhang, M.Y., Ersal, T., Yang, X.J.: Effects of a delay compensation aid on teleoperation of unmanned ground vehicles. In: Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction, pp. 179–180. Association for Computing Machinery, Chicago, USA (2018). https://doi.org/10.1145/3173386.3177064
Mathan, S., Hyndman, A., Fischer, K., Blatz, J., Brams, D.: Efficacy of a predictive display, steering device, and vehicle body representation in the operation of a lunar vehicle. In: Conference Companion on Human Factors in Computing Systems, pp. 71–72. Association for Computing Machinery, Vancouver, BC, Canada (1996)
Acknowledgements
The authors acknowledge the contribution of the survey participants who took part to evaluate the teleoperation system and provided valuable feedback. The authors also thank Mr. Hassan Mahmood and Mr. Shinsuke Matsubara for their suggestions and help with Matlab and Simulink as well as Dr Guy Gallasch and Dr Robert Hunjet from DSTG for their feedback on the teleoperation simulation and enhancement model.
Funding
Open Access funding enabled and organized by CAUL and its Member Institutions This research has been funded by the ECU-DSTG Industry PhD scholarship.
Author information
Authors and Affiliations
Contributions
The contributions of the authors are as follows.
• MD Moniruzzaman: Conceptualisation, model creation, data collection, experimentation, manuscript preparation.
• Alexander Rassau: Conceptualisation, supervision, manuscript preparation, and editing.
• Douglas Chai: Conceptualisation, supervision, manuscript preparation, and editing.
• Syed Mohammed Shamsul Islam: Conceptualisation, supervision, manuscript preparation, and editing.
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Ethics approval
Appropriate approval has been granted by the Human Ethics Committee (HREC) of Edith Cowan University (ECU) prior to any operator data collection. Approval number 2020-01643.
Consent to participate
The authors declare that appropriate consent has been taken from the participants of the human survey described in this paper.
Consent for publication
This paper does not contain materials, figures, tables, etc that have been published anywhere else. Therefore consent for publication is not applicable.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix: Incorporating Control Input Delay with Visual Latency
Appendix: Incorporating Control Input Delay with Visual Latency
Latency into the simulation platform can be achieved either by adding delay to the visual feed or by delaying the control input to the simulation platform. The authors of the paper are confident that the impacts of either of these delay factors, or a combination of them, are perceived as the same from the perspective of a teleoperator. This appendix discusses the methods through which the control input delay can be achieved and the impacts on a teleoperator.
In the main sections of the paper, latency was achieved purely by delaying the visual feed. To achieve the visual delay, a single computer unit was used to connect the simulating platform with controller devices, visual feed receiver camera, and the latency control and teleoperation enhancement Simulink® model. However, to incorporate control input delay along with the visual delay, we have used another personal computer (PC) unit, referred to as PC-2 in the rest of this section. The system diagram of the modified teleoperation simulation system has been provided in Fig. 13. In this modified system, the controller devices (steering wheel, brake and acceleration pedals) are connected to the PC-2. The vehicle simulation platform (game engine), the visual feed receiver, and teleoperation enhancement model is hosted by the previously used PC, named PC-1 hereafter. The two computers can be connected either by local area network (LAN), wide area network (WAN), or Internet connection. For our case, the computers were connected to the university LAN.
Aside from these new amendments, the rest of the system is as described in Section 3.
To add latency between the two computers we have used a third party software called Clumsy. Clumsy uses the Windows Packet Divert library to stop, capture, lag, drop, or tamper with packets on a living network. Any amount of system to system latency is achievable using this tool. An example of varied latency between two computers using Clumsy is shown in Fig. 14. To connect and receive control input from PC-2 to PC-1 we have used another third party software called VirtualHere. Although USB devices usually need to be directly connected to a computer to be used, VirtualHere facilitates the transmission of USB signals over a LAN, WAN or Internet connection to a remote machine, allowing for virtual connection of USB devices over a network.
To investigate the impact of control latency we have run teleoperation sessions on both on-road and off-road tracks. These are the same tracks used by the survey participants. We have compared the outcomes with the non-assisted delayed visual feed outcomes by the participants (From Table 3). The non-assisted delayed feed is affected by the visual delay only. The comparison is in Table 5 below. For both the on-road and off-road tracks, the average speed and time taken to complete one lap is very similar. A small amount of difference exists as the control only delay is a little higher than that of the visual delay.
Table 5 demonstrates that a certain amount of latency or delay in the teleoperation control loop, regardless of its source, will have the same impact on the teleoperator. To reinforce this statement and to prove the robustness of our teleoperation enhancement technique, we have conducted teleoperation runs on on-road and off-road tracks where the total delay in the loop consists of both control input delay and visual output delay. The outcome of the teleoperation sessions is provided in Table 6. For the on-road teleoperation session, the cumulative latency was 1200 ms where the control input latency was 560 ms and the visual feed latency was 640 ms. Without any video transformation based assistance it took 499 s to complete the 2 km track with an average speed of 14.42 km/h. Using our sliding window enhancement, the completion time reduced to 428 s with an average speed of 16.8 km/h. Using the sliding and zooming window both the time and speed improved further. Incidents of oscillation and out of track events also improved substantially using our assisted windows.
For the off-road track, the figures correspond with the previous numbers. In this case, the total latency is 1076 ms where 560 ms latency is induced from the controller to PC-2 and 516 ms latency is induced by our Simulink® visual transformation model. Using the sliding window the completion time reduced from 385 s to 282 s. For sliding and zooming the time is further reduced. The oscillation and out of track events reduced to zero for both the sliding only and sliding with zooming windows. Table 6 reflects the same outcome to that obtained via the participants’ survey. It reconfirms the findings of our research that our video transformation-based assisted windows enhance an operator’s performance for high latency ground-vehicle teleoperation. Based on these results, it can confidently be stated that whether the delay is present only in the visual feed, only in the control input, or a combination of the two (as would be the case in a real-world scenario) the impact is effectively the same from the operator perspective. This validates the platform that has been setup to evaluate the teleoperation enhancement techniques.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Moniruzzaman, M., Rassau, A., Chai, D. et al. High Latency Unmanned Ground Vehicle Teleoperation Enhancement by Presentation of Estimated Future through Video Transformation. J Intell Robot Syst 106, 48 (2022). https://doi.org/10.1007/s10846-022-01749-3
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10846-022-01749-3