Securing Trustworthy XR Interactions through Human-Machine Networks in Healthcare
Description
Securing trustworthy XR interactions is a core requirement for the Future of Work that uses immersive technologies and mixed reality. Our specific context includes applications of remote collaboration between VR and AR users in a telemedicine context. For example, we are interested in exploring how users trust the use of mixed-reality devices for this type of application, including current medical students who will be at the forefront of the integration of immersive tech in medicine in the future.
Securing trust is critical for adoption, and even more critical in a medical context when it comes to data security and privacy. This is integrated into our project by investigating the role authentication plays for connecting with remote medical experts, and whether users trust the instructions or medical diagnoses they may receive from remote caregivers in this system. We will be developing a VR/AR remote collaboration system as part of this project. Methods we plan to explore for increasing trust in this ecosystem include authentication methods and expert evaluation of the anticipated data flows to identify areas that need enhanced security as a result of increased sensors included on mixed-reality devices and the applications they are likely to support in tele-medicine.
Documentation
Reports
The primary research activities conducted during Fall were broken up into literature review of XR collaborative systems in a medical context, DepthKit Studio configuration and collaboration with
University Libraries/ARIES, system exploration for alternative collaboration systems.
Literature Review
Literature review was conducted of research results and existing systems that fit our collaboration model (local user and remote expert). Papers were first characterized based on the collaboration configuration, the data transferred, and context.
We then synthesized new labels for the most common sensors/systems to examine trade-offs between feasibility in deployment, user experience, and potential privacy risks.
DepthKit Studio Testing
The project team initiated a collaboration with ARIES and the University Libraries to start the configuration of tele-operation system for VR/AR. This was inspired by prior work from Cornell Tech published at an IEEE VR workshop.
However, during this process we encountered many roadblocks in terms of calibration, capture of the user and the environment in one run, networking captured data to a nearby computer, and compatibility of the system with a Magic Leap 2 AR headset.
System Exploration
Beyond DepthKit Studio, we reviewed other tele-operation systems for VR/AR, including Magic Leap’s Remote Assist and Workshop software, MakeSEA Catapult, and building a tele-operation networked application from scratch using Unity and 360-degree videos. From this evaluation, we determined that MakeSEA Catapult and Magic Leap Remote Assist are the most viable options. Additionally, we explored the potential to integrate generative AI for the target task through Microsoft’s CoPilot AI, which offers to provide automated annotations and responses for frontline workers using AR headsets. So far, we have not been able to identify a useable system for this situation, but plan to explore this recent breakthrough in AI in the context of remote healthcare when using spatial computing/mixed reality devices.
Spring 2024 Plans
Based on our results, we have determined a modified plan for Spring 2024. First, we will conduct up to two design or focus group workshops on the context of emerging technologies, in line with the original proposal. The only modification to these workshops would be related to a more interactive workshop centered around and using generative AI mechanisms (language and image synthesis models). Second, we narrowed our scope for the preliminary remote user study to make use of Magic Leap 2 Remote Assist, as the plug in play solution allows us to execute a user study with three clear conditions, no remote assist (baseline), remote assist using only the first-person view from the local client, and remote assist using both the first-person view and the third-person view enabled by the environment sensing built into the AR device. The user study can then evaluate the ability of remote experts (medical students) to guide novices (general participants from the VT community) through the remote medical task to evaluate usability and privacy concerns. Based on our literature review, this setting is representative of a practical and feasible deployment to areas with limited access, as it only involves sending a mobile lightweight headset to the client instead of a ten sensor DepthKit Studio set up that requires precise calibration. The user study is in line with the originally proposed work, only remove the need to require the remote expert to use a VR headset and instead fall back to the use of a traditional desktop setting for guidance, and includes the tools and gestures originally proposed from a VR remote expert and the produced risks to viewing the local client’s home environment during guidance.