Technical Analysis

A Paradigm Shift in
Browser-Based Rendering

The Uruk Neural Renderer replaces traditional physics-based lighting with a trained neural network, delivering cinematic visuals directly in WebGPU at 60 FPS.

0M
Parameters
Runtime model size
0 FPS
Frame Rate
On mid-range hardware
0ms
Inference
Per-frame latency
0GB
Download
No baked assets needed
The Problem

WebGPU Has a Quality Ceiling

Despite being a massive leap from WebGL, standard WebGPU rendering operates under severe constraints compared to native desktop APIs. The result is a visual quality roughly equivalent to late PS3 or early PS4 era — clean but lifeless.

No Hardware Ray Tracing

WebGPU lacks access to RT cores. All ray tracing must be done in software via compute shaders, which is prohibitively slow for complex scenes.

Texture & Memory Limits

Strict limits on read-write storage textures. No async compute, no mesh shaders, no variable rate shading. Maximum 16 textures per shader.

Uniformity Constraints

Strict uniformity analysis makes complex branching in shaders difficult to optimize, limiting advanced rendering techniques.

How It Works

The Neural Rendering Pipeline

Instead of simulating millions of light ray intersections, Uruk predicts the final cinematic frame from cheap G-buffer inputs using a trained neural network.

Neural rendering pipeline: G-buffers flow through neural network to produce cinematic output
01

G-Buffer Extraction

The game engine outputs cheap, unlit geometry data: depth, normals, albedo, and material properties. This takes only 2-3ms.

02

Neural Inference

A 2M-parameter neural network processes the G-buffers in a single forward pass, predicting cinematic lighting, GI, reflections, and atmosphere.

03

Cinematic Output

The final frame features dynamic global illumination, accurate reflections, volumetric effects, and cinematic color grading — all from a single inference step.

Visual Comparison

Standard WebGPU vs. Uruk Neural

Drag the slider to compare standard WebGPU rasterization against Uruk's neural rendering. Notice the global illumination, volumetric light, and accurate reflections.

Uruk Neural Renderer output
Standard WebGPU output
Standard WebGPU
Uruk Neural

Standard WebGPU

  • Flat, baked or screen-space lighting only
  • No global illumination or light bouncing
  • Screen-space reflections cut off at edges

Uruk Neural Renderer

  • Dynamic global illumination with color bleeding
  • Volumetric god rays and atmospheric scattering
  • Full-scene reflections via neural prediction
Performance Analysis

Frame Budget Breakdown

UE5 spends its frame budget across 6+ rendering passes. Uruk collapses this into 3 passes, with the neural inference replacing Lumen GI, reflections, and shadow computation.

Unreal Engine 5

Lumen + Nanite + VSM

18ms

per frame

Base Pass (G-buffer)
3.5ms
Nanite Rasterization
2.5ms
Lumen GI
4.5ms
Lumen Reflections
2.5ms
Virtual Shadow Maps
2.5ms
Post-Processing (TSR)
2.5ms

Uruk Neural Renderer

Neural Inference Pipeline

14ms

per frame

G-buffer Extraction
2.5ms
Neural Inference
10ms
Temporal Blending
1.5ms

22% faster total frame time — and runs in a browser

Quantitative Analysis

Head-to-Head Comparison

MetricStandard WebGPUUnreal Engine 5Uruk Neural
Global IlluminationBaked / Screen-spaceDynamic, Ray-tracedDynamic, Neural Predicted
ReflectionsScreen-space onlyRay-traced (Expensive)Neural Predicted
Lighting Cost (1080p)~2-4 ms~6-8 ms~8-12 ms
Hardware RT RequiredNot SupportedYes (Best Quality)No
Asset FootprintMedium (Baked maps)Massive (Gigabytes)Tiny (No baked data)
Platform TargetBrowserDesktop / ConsoleBrowser & Native
The Secret Weapon

Knowledge Distillation

We train a massive 42M-parameter Teacher model on NVIDIA B200 GPUs to deeply understand cinematic lighting. Then we distill that knowledge into a tiny 2M-parameter Student model that ships to the user.

0M
Teacher Parameters
0M
Student Parameters

Deployment Targets

GPU
RTX 4080
<8ms
Apple
M3 Max (CoreML)
<12ms
Web
WebGPU (Browser)
<25ms
Knowledge distillation: 42M Teacher model distilled into 2M Student model for WebGPU, CoreML, and ONNX deployment
Live Training Metrics

Training on NVIDIA B200

60 epochs at 720p with cosine warm-restart scheduling. The model learns to map G-buffers to cinematic frames using L1, perceptual, and temporal consistency losses.

0.0dB
PSNR
0.000
SSIM
0.000
Train Loss

Loss Curves

TrainVal
Epoch 1Epoch 60

PSNR Over Time

32.1 dB
Epoch 1Epoch 60

Epoch Timeline

60 / 60
WarmupCosine Restart 1Cosine Restart 2Convergence
GPU
B200
180GB VRAM
Resolution
720p
1280 x 720
Batch Size
4
bf16 mixed precision
Est. Time
~55 min
$4.99/hr
Interactive Demo

Neural Rendering in Your Browser

Once the model is trained and exported, this section will run the Uruk Neural Renderer live via ONNX Runtime Web and WebGPU — no downloads, no plugins, just cinematic rendering.

WebGPU Neural Renderer

Model training in progress — demo activates after export

-- FPS
ONNX Runtime Web + WebGPU EP
Model
2M params
Input
720p G-buffers
Output
720p + 2x upscale
Runtime
ONNX WebGPU
Target
< 12ms / frame
Vision — The Roadmap

Text to Neural World

Type a prompt. Get a playable, cinematically lit 3D world in the browser. No external APIs, no vendor lock-in — every model weight is yours.

uruk-engine prompt
>

"Ancient Mesopotamian city at sunset with sandstorms and torch-lit bazaars, ziggurats casting long shadows, canal reflections, bustling marketplace"

Scene Graph~2s
Mesh Gen~8s
Materials~5s
Layout~1s
Neural Render~0.01s/frame

Five Custom Neural Models

UrukScenePlanner

Parses text prompts into structured scene graphs with spatial relationships

Llama 3.1 8B fine-tune~8 hrs

UrukMeshGen

Generates 3D meshes from text descriptions with LODs and collision

InstantMesh / OpenLRM~48 hrs

UrukMaterialNet

Creates PBR materials and textures from text descriptions

Stable Diffusion XL fine-tune~24 hrs

UrukAtmosphereNet

Foundation neural renderer trained on 50+ diverse scene styles

Uruk Renderer (ours)~100 hrs

UrukLayoutNet

Spatial reasoning transformer for object placement and scene composition

Custom transformer~12 hrs
~$960

Total training cost for the complete text-to-world neural stack

~192 GPU hours on B200

100% Open-Source Training Data

SourceContentLicenseUse
Objaverse800K+ 3D modelsCC-BYMesh training
ShapeNet51K modelsResearchMesh topology
Ambient CGPBR materialsCC0Material training
Poly HavenHDRIs + texturesCC0Lighting
Structured3DIndoor layoutsOpenScene layout
Scene FactoryCustom generatedProprietaryAll training

Full Ownership

Every model weight, every training pipeline, every inference runtime. No API calls, no per-generation fees, no vendor dependencies.

Compounding Moat

Each new scene trained makes the foundation model better. Training data + model weights compound over time — a defensible technical advantage.

Browser-First

The only text-to-world pipeline that runs entirely in WebGPU. Instant play, zero download, cinematic quality — in any browser.

ESTIMATED TIMELINE

M1-2
Scene Planner
M3-6
Asset Pipeline
M4-6
Foundation Renderer
M7-8
Scene Assembly
M9
Demo
The Material Differences

A Different Rendering Paradigm

By shifting the computational burden from runtime physics simulation to ahead-of-time neural training, Uruk enables AAA-quality visuals directly in the web browser.

Quality vs. WebGPU

Uruk delivers UE5-level lighting — bounced light, color bleeding, accurate reflections — in the browser. Standard WebGPU cannot achieve this dynamically because it lacks hardware ray tracing. Uruk bridges this gap by replacing ray tracing with neural inference.

Performance vs. Unreal Engine

While UE5 spends 6-8ms on Lumen GI and reflections, Uruk achieves comparable quality in ~8-12ms — without requiring massive engine overhead, complex acceleration structures, or gigabytes of VRAM.

The Distillation Pipeline

The B200 training pipeline is the secret weapon. A 42M-parameter Teacher learns cinematic lighting, then distills into a 2M-parameter Student. This tiny model downloads instantly and runs at 60 FPS in WebGPU.

"Users get the visual fidelity of a 50GB console game in an instant-loading browser tab. Developers get a streamlined art pipeline where they only need to provide basic geometry and materials, letting the AI handle the cinematic lighting."