In partnership with TELUS, and the University of Calgary, this project focused on designing and prototyping an end-to-end XR platform where extended reality (XR) components such as video processing and XR rendering are deployed as microservices across core and edge clouds. The aim was to enhance Wi-Fi throughput visualization and optimization in indoor environments using Digital Twin (DT) technology, and remote support solution to place motion sensor.
The project introduces a user-friendly XR solution where a Digital Twin of a home can be generated using LiDAR and ARKit, throughput can be measured at specific points, and performance mapped in 3D. This approach not only improves technical accuracy but also provides a direct benefit to TELUS customers: helping them identify optimal Wi-Fi router placements, reducing negative feedback, and enhancing brand loyalty.
Project Objectives
Objective 1: Digital Twin Creation
Scan indoor environments using LiDAR and ARKit (Apple RoomPlan).
Generate 3D floor plans with dimensions and furniture recognition.
Export DTs in USD/USDZ formats for compatibility with design tools.
Objective 2: Throughput Measurement & Visualization
Shift focus from traditional signal strength to throughput-based assessment, mitigating environmental and device variability.
Measure download, upload, and latency via APIs (e.g., SpeedOf.Me) and Unity assets.
Visualize results using an innovative bubble-based method:
Green = High (>100 Mbps)
Yellow = Medium (50–100 Mbps)
Red = Low (<50 Mbps)
Provide intuitive, real-time visual feedback in immersive environments.
Objective 3: Enhancing User Experience
Develop interactive interfaces for selecting test locations.
Implement dollhouse and overlaid DT visualizations for better context.
Improve usability via color-coding and simple UI cues.
Conduct pilot and user studies to evaluate interpretation and effectiveness.
Objective 4: Collaboration & Industry Relevance
Deliver economic benefits by reducing customer complaints and optimizing service adoption for TELUS.
Create scalable XR microservices deployable over 5G networks.
Provide research insights into HCI, immersive analytics, AR/VR, and AI integration.
Objective 5: Motion Sensor Placement with HMD Integration
Enable users to scan their home environments using an iPad equipped with LiDAR.
Recreate the scanned environment inside a head-mounted display (HMD) for immersive visualization.
Provide real-time guidance for optimal motion sensor placement across different rooms.
Allow users to preview coverage areas, blind spots, and sensor effectiveness before installation.
Improve home automation and security workflows by combining Digital Twin modeling with immersive XR simulations.
Methodology & Implementation
Phase 1: Focused on DT creation using Apple LiDAR, throughput measurement via third-party APIs, and dollhouse-style visualization.
Phase 2: Integrated Unity assets for streamlined scanning, throughput testing, and AR overlays with enhanced UI design.
Phase 3 : Extend DT workflows into HMD-based immersive environments, supporting interactive sensor-placement planning.
University of Calgary – SEER Lab and Ethereal Lab
Parisa Daeijavad (Ph.D. studen)
Naman Bhoj (MSc student)
Nami Modarressi (MSc student)
- Ali Shahidi (MSc student)
- Farhan Aslam (MSc Student)
Supervisors: Dr. Frank Maurer , Dr. Richard Zhao, Dr. Diwakar Krishnamurthy