Accelerating UAE’s Shift to Smart Mobility
Building the Benchmark for Autonomous UAE
We provide the open-source benchmark dataset and expert TaaS (Testing-as-a-Service) to help global OEMs achieve safe, certified deployment.
We provide the open-source benchmark dataset and expert TaaS (Testing-as-a-Service) to help global OEMs achieve safe, certified deployment.
We provide the open-source benchmark dataset and expert TaaS (Testing-as-a-Service) to help global OEMs achieve safe, certified deployment.



Autonomous Mobility
Tele-Operations
Data Infrastructure
AI & Simulation
Safety Systems
5G Connectivity
Smart Cities
Fleet Management
Mapping & Localization
Edge Computing
Autonomous Mobility
Tele-Operations
Data Infrastructure
AI & Simulation
Safety Systems
5G Connectivity
Smart Cities
Fleet Management
Mapping & Localization
Edge Computing
Autonomous Mobility
Tele-Operations
Data Infrastructure
AI & Simulation
Safety Systems
5G Connectivity
Smart Cities
Fleet Management
Mapping & Localization
Edge Computing
Local Data Gap
UAE roadways are heavily underrepresented in global AV datasets. Even with regional efforts, less than 1% of public autonomous-vehicle data reflects UAE-specific conditions, limiting the reliability of imported perception models and slowing national deployment.
Local Data Gap
UAE roadways are heavily underrepresented in global AV datasets. Even with regional efforts, less than 1% of public autonomous-vehicle data reflects UAE-specific conditions, limiting the reliability of imported perception models and slowing national deployment.
Local Data Gap
UAE roadways are heavily underrepresented in global AV datasets. Even with regional efforts, less than 1% of public autonomous-vehicle data reflects UAE-specific conditions, limiting the reliability of imported perception models and slowing national deployment.
Environmental Challenges
Desert climate reduces sensor performance by up to 65%. Heat shimmer, glare, sand dust, and low-texture desert scenes degrade LiDAR and camera reliability—factors mostly absent from U.S. and European datasets, causing perception and tracking errors during UAE testing.
Environmental Challenges
Desert climate reduces sensor performance by up to 65%. Heat shimmer, glare, sand dust, and low-texture desert scenes degrade LiDAR and camera reliability—factors mostly absent from U.S. and European datasets, causing perception and tracking errors during UAE testing.
Environmental Challenges
Desert climate reduces sensor performance by up to 65%. Heat shimmer, glare, sand dust, and low-texture desert scenes degrade LiDAR and camera reliability—factors mostly absent from U.S. and European datasets, causing perception and tracking errors during UAE testing.
Infrastructure Variation
Over 120 distinct roadway and signage variations across the UAE. Differences in road markings, lane geometry, lighting conditions, and signage standards make it difficult for imported autonomy stacks to adapt without extensive local validation and re-training.
Infrastructure Variation
Over 120 distinct roadway and signage variations across the UAE. Differences in road markings, lane geometry, lighting conditions, and signage standards make it difficult for imported autonomy stacks to adapt without extensive local validation and re-training.
Infrastructure Variation
Over 120 distinct roadway and signage variations across the UAE. Differences in road markings, lane geometry, lighting conditions, and signage standards make it difficult for imported autonomy stacks to adapt without extensive local validation and re-training.
Our Vision
Vehicle + sensor overlay (engineering drawing on hover)
This view represents the physical validation platform used to design and test production-grade perception stacks. The engineering overlay illustrates the placement of LiDAR, radar, and camera sensors, reflecting real mounting constraints, sensor overlap, and field-of-view considerations required for accurate multi-modal sensor fusion in real-world environments.
Remote driving of CARLA over the internet (MVP)
This is a screen capture from our active MVP demonstrating real-time teleoperation of a CARLA vehicle rendered on a remote GPU machine. The prototype validates end-to-end control flow, network latency, video streaming, and telemetry exchange using a browser-based operator interface, serving as a testbed for remote intervention, data acquisition, and validation workflows.
Vehicle + sensor overlay (engineering drawing on hover)
This view represents the physical validation platform used to design and test production-grade perception stacks. The engineering overlay illustrates the placement of LiDAR, radar, and camera sensors, reflecting real mounting constraints, sensor overlap, and field-of-view considerations required for accurate multi-modal sensor fusion in real-world environments.
Data view → real-world driving (AI transition)
This visualization illustrates our long-term vision of bridging structured perception data, simulation, and on-road autonomous driving. It represents how validated sensor data and scenario knowledge transition from data-centric views into real-world vehicle behavior through domain-adaptive perception, simulation-driven testing, and deployment-ready validation pipelines.
Remote driving of CARLA over the internet (MVP)
This is a screen capture from our active MVP demonstrating real-time teleoperation of a CARLA vehicle rendered on a remote GPU machine. The prototype validates end-to-end control flow, network latency, video streaming, and telemetry exchange using a browser-based operator interface, serving as a testbed for remote intervention, data acquisition, and validation workflows.
Vehicle + sensor overlay (engineering drawing on hover)
This view represents the physical validation platform used to design and test production-grade perception stacks. The engineering overlay illustrates the placement of LiDAR, radar, and camera sensors, reflecting real mounting constraints, sensor overlap, and field-of-view considerations required for accurate multi-modal sensor fusion in real-world environments.
Vehicle + sensor overlay (engineering drawing on hover)
This view represents the physical validation platform used to design and test production-grade perception stacks. The engineering overlay illustrates the placement of LiDAR, radar, and camera sensors, reflecting real mounting constraints, sensor overlap, and field-of-view considerations required for accurate multi-modal sensor fusion in real-world environments.
Our Vision
Vehicle + sensor overlay (engineering drawing on hover)
This view represents the physical validation platform used to design and test production-grade perception stacks. The engineering overlay illustrates the placement of LiDAR, radar, and camera sensors, reflecting real mounting constraints, sensor overlap, and field-of-view considerations required for accurate multi-modal sensor fusion in real-world environments.
Vehicle + sensor overlay (engineering drawing on hover)
This view represents the physical validation platform used to design and test production-grade perception stacks. The engineering overlay illustrates the placement of LiDAR, radar, and camera sensors, reflecting real mounting constraints, sensor overlap, and field-of-view considerations required for accurate multi-modal sensor fusion in real-world environments.
Remote driving of CARLA over the internet (MVP)
This is a screen capture from our active MVP demonstrating real-time teleoperation of a CARLA vehicle rendered on a remote GPU machine. The prototype validates end-to-end control flow, network latency, video streaming, and telemetry exchange using a browser-based operator interface, serving as a testbed for remote intervention, data acquisition, and validation workflows.
Remote driving of CARLA over the internet (MVP)
This is a screen capture from our active MVP demonstrating real-time teleoperation of a CARLA vehicle rendered on a remote GPU machine. The prototype validates end-to-end control flow, network latency, video streaming, and telemetry exchange using a browser-based operator interface, serving as a testbed for remote intervention, data acquisition, and validation workflows.
Data view → real-world driving (AI transition)
This visualization illustrates our long-term vision of bridging structured perception data, simulation, and on-road autonomous driving. It represents how validated sensor data and scenario knowledge transition from data-centric views into real-world vehicle behavior through domain-adaptive perception, simulation-driven testing, and deployment-ready validation pipelines.
Data view → real-world driving (AI transition)
This visualization illustrates our long-term vision of bridging structured perception data, simulation, and on-road autonomous driving. It represents how validated sensor data and scenario knowledge transition from data-centric views into real-world vehicle behavior through domain-adaptive perception, simulation-driven testing, and deployment-ready validation pipelines.
Solution
What We Offer
The three-layer technical stack powering safe, scalable autonomous mobility across the UAE.
About the Founder
Engineering the UAE’s Next-Generation Autonomous Mobility Infrastructure
Y+
Y+
Y+
5+ years across autonomous systems engineering, V&V, simulation, and tele-operations.
With a background in autonomous vehicles, tele-operations, and large-scale mobility testing, my work integrates safety-critical V&V frameworks, scenario-based simulation, tele-operation control architectures, and high-resolution data pipelines. This includes system-level validation across perception, planning, and control; certification-grade testing under TÜV SÜD methodology; and the development of tele-driving workflows capable of supporting remote operations at national scale.
By combining simulation-driven verification, real-world test campaigns, and structured data workflows, I focus on building the technical foundation required for reliable, scalable, and regulation-aligned autonomous mobility deployment across the UAE.
Who we are & what we do
(The Mission) We are the UAE's strategic partner for AV validation. We provide the national benchmark dataset and expert TaaS (Testing-as-a-Service) to get global OEMs certified and on the road, safely.
Who we are & what we do
(The Mission) We are the UAE's strategic partner for AV validation. We provide the national benchmark dataset and expert TaaS (Testing-as-a-Service) to get global OEMs certified and on the road, safely.
Who we are & what we do
(The Mission) We are the UAE's strategic partner for AV validation. We provide the national benchmark dataset and expert TaaS (Testing-as-a-Service) to get global OEMs certified and on the road, safely.
What we build
(The Funnel Model) Our "open-core" platform for market entry: 01 A free, public Benchmark Dataset. 02 A paid TaaS & Tele-Operation Service. 03 A paid Digital Twin & Scenario Library.
What we build
(The Funnel Model) Our "open-core" platform for market entry: 01 A free, public Benchmark Dataset. 02 A paid TaaS & Tele-Operation Service. 03 A paid Digital Twin & Scenario Library.
What we build
(The Funnel Model) Our "open-core" platform for market entry: 01 A free, public Benchmark Dataset. 02 A paid TaaS & Tele-Operation Service. 03 A paid Digital Twin & Scenario Library.
Partner with us
(The Target) We collaborate with government bodies (RTA, ITC/SAVI) to set the standard, and with global OEMs to provide the validation services they need to enter the UAE market.
Partner with us
(The Target) We collaborate with government bodies (RTA, ITC/SAVI) to set the standard, and with global OEMs to provide the validation services they need to enter the UAE market.
Partner with us
(The Target) We collaborate with government bodies (RTA, ITC/SAVI) to set the standard, and with global OEMs to provide the validation services they need to enter the UAE market.












Invest in the UAE's AV Enabler.
Let's Build Something
Amazing Together
We are building the open-source benchmark and TaaS platform to unlock the market for global OEMs. This is the foundational investment to accelerate UAE Vision 2031. See the full plan.





