When I joined Metamorphic Design Company, one of my first briefs was brutal in its simplicity: clients need to see their space before we build it, and they need to see it convincingly. Not a flat PDF. Not a cartoonish render. Something that makes them feel they are already standing in the finished interior.
What I ended up shipping was a full production pipeline combining 8K photoreal renders, VR walkthroughs, and on-site AR measurement using the Aria Gen 2 glasses. Here is how it came together.
The Three Problems We Had To Solve
- Production-grade renders: standard architectural visualization tools produce 1080p or 4K output. For premium villas in Dubai, that is not enough. Clients zoom in on details.
- Real-time immersion: a static render is a picture. Clients want to walk around, look up at the ceiling, open doors, change finishes in the moment.
- On-site precision: before the render matters, our site supervisors need to capture exact measurements from the physical space, and they need to do it without dragging a laser measure around for three hours.
Part 1: The 8K Render Pipeline
We standardized on an Unreal Engine 5 based workflow with Lumen global illumination and Nanite virtualized geometry. Textures come from a proprietary material library we built over six months of scans. Everything feeds into a Python orchestration layer that handles:
- Asset ingestion from the design team's shared drive
- Scene assembly based on Pipedrive project metadata
- Distributed render farm submission (custom scheduler on top of Kubernetes)
- Automatic LoD generation for the VR viewer
- Delivery to the client portal with versioning
The hard part was the render farm. Each 8K frame takes 40 to 90 seconds on a single RTX 4090. A villa walkthrough has thousands of frames. We run 24 GPU nodes in parallel, with a priority queue so premium clients jump the line.
Part 2: The VR Viewer
The render farm produces beautiful stills, but stills are not experiences. For walkthroughs, we built a lightweight Three.js based viewer that streams mesh LoDs and lightmaps based on headset pose.
Key choices:
- Three.js over Unity for the viewer: no install required, runs in a browser, works on Quest and Vision Pro out of the box
- glTF 2.0 with Draco compression: single industry standard, good tool chain
- PBR materials with ACES tone mapping: the renders look cinematic without extra post-processing
- Hot-swap finish variations: clients can toggle between three flooring options in real time without a reload
The first version choked on villa-scale scenes. We hit 3 frames per second on a Quest 3 with the full mesh loaded. The fix was aggressive LoD switching plus occlusion culling. Now we run at a steady 72 fps even on the largest scenes.
Part 3: Aria Gen 2 AR Glass for On-Site Work
This is the part I am proudest of. Meta's Aria Gen 2 research glasses combine depth sensors, SLAM, and passthrough AR in a form factor that site supervisors can actually wear for an hour at a time. We integrated them into our workflow for three specific jobs:
1. Measurement Capture
Our site supervisor walks into an empty villa, points at corners, and the glasses capture exact distances, floor plans, and ceiling heights. The data flows straight into our BOQ (Bill of Quantities) system. A three hour manual measurement job now takes 20 minutes.
2. Cost Estimate Overlay
Procurement walks through the same space wearing the glasses and sees an AR overlay of running material costs per room. "This marble is 280 AED per square meter, total for this floor is 48,000 AED." Decisions get made on site instead of in follow-up meetings a week later.
3. Client Design Review
The killer use case. The client stands in the half-built villa and sees the final design overlaid on the real walls through the glasses. They see the couch, the lighting, the artwork, exactly where it will go. This has closed deals that would have stalled for weeks.
The Tech Stack
Content Pipeline:
Unreal Engine 5 + Lumen + Nanite
Python orchestrator (FastAPI)
Custom Kubernetes render farm (24 GPUs)
pgvector for asset similarity search
VR Viewer:
Three.js with WebXR
glTF 2.0 with Draco
Custom LoD streaming
PBR + ACES tone mapping
AR Integration:
Aria Gen 2 SDK
SLAM and depth integration
WebRTC for live client viewing
Direct integration with MetaForge ERP
What I Learned
Fidelity is a product metric. Clients paying for a 15 million AED villa can tell the difference between 4K and 8K. Skipping that last step to save render time is false economy.
The viewer is more important than the render. A perfect render in a clunky viewer feels worse than a slightly lower quality render in a smooth viewer. Interaction latency is everything.
AR beats VR for on-site work. VR is great when you are sitting in an office. On a construction site, you need to see the real world plus overlays, not a fully synthetic environment.
The best visualization pipeline is invisible. The client never asks about the tech. They just see their home.
Building this pipeline was the most cross-disciplinary project of my career so far. It touched computer graphics, distributed systems, ML, SLAM, and product design. If you are working at the intersection of AI and 3D, the next five years are going to be wild.
