diff options
Diffstat (limited to 'tools/cnn_v2_test')
| -rw-r--r-- | tools/cnn_v2_test/README.md | 251 | ||||
| -rw-r--r-- | tools/cnn_v2_test/index.html | 2049 |
2 files changed, 0 insertions, 2300 deletions
diff --git a/tools/cnn_v2_test/README.md b/tools/cnn_v2_test/README.md deleted file mode 100644 index d41a00f..0000000 --- a/tools/cnn_v2_test/README.md +++ /dev/null @@ -1,251 +0,0 @@ -# CNN v2 Testing Tool - -WebGPU-based browser tool for testing trained CNN v2 weights. - ---- - -## Features - -- Drag-drop PNG images and `.bin` weights (or click to browse) -- Real-time CNN inference with WebGPU compute shaders -- View modes: CNN output, original input, difference (×10) -- Adjustable blend amount and depth -- Data-driven pipeline (supports variable layer count) -- GPU timing display -- **Left Panel:** Weights info + kernel visualization (1px/weight, all layers) -- **Right Panel:** Layer activation viewer with 4-channel split + 4× zoom - ---- - -## Requirements - -- Browser with WebGPU support: - - Chrome/Edge 113+ (enable `chrome://flags/#enable-unsafe-webgpu` if needed) - - Safari 18+ (macOS Ventura+) -- Trained CNN v2 weights in binary format (`.bin`) -- Test images (PNG format) - ---- - -## Usage - -### 1. Open Tool - -```bash -open tools/cnn_v2_test/index.html -``` - -Or use a local server to avoid CORS: -```bash -python3 -m http.server 8000 -# Open http://localhost:8000/tools/cnn_v2_test/ -``` - -### 2. Load Data - -1. **Drop `.bin` weights** into left sidebar zone (or click to browse) -2. **Drop PNG image** anywhere in center canvas area -3. CNN runs automatically when both loaded - -### 3. Layout - -**Left Sidebar:** -- Weights drop zone (click or drag-drop `.bin` files) -- Weights info panel (layer specs, ranges, file size) -- Weights visualization (click Layer 0/1/2 buttons) - - 1 pixel per weight, all input channels horizontally - - Output channels (Out 0-3) stacked vertically - -**Center Canvas:** -- Main output view (CNN result, original, or diff) -- Keyboard: `SPACE` = original, `D` = diff (×10) - -**Right Sidebar:** -- Layer selection buttons (Static 0-3/4-7, Layer 0/1/2) -- 4 small activation views (Ch0/1/2/3) in a row -- Large zoom view below (4× magnification, follows mouse) - -**Header Controls:** -- **Blend:** Mix between original (0.0) and CNN output (1.0) -- **Depth:** Uniform depth value for all pixels (0.0–1.0) -- **View:** Current display mode - -**Footer:** -- Status: GPU timing (ms), image dimensions, view mode -- Console: Timestamped event log (file loads, errors) - ---- - -## Preparing Test Data - -### Export Weights - -```bash -# From trained checkpoint -./training/export_cnn_v2_weights.py \ - checkpoints/checkpoint_epoch_100.pth \ - --output-weights tools/cnn_v2_test/test_weights.bin -``` - -Binary format: 16-byte header + 20 bytes per layer + f16 weights (~3.2 KB for 3-layer model) - -### Test Images - -Use training images or any PNG: -```bash -# Copy test image -cp training/input/test.png tools/cnn_v2_test/ -``` - -**Note:** Grayscale images automatically converted to RGB. - ---- - -## Validation - -### Visual Comparison - -Compare browser output with C++ tool: - -```bash -# Generate C++ output -./build/cnn_test training/input/test.png /tmp/cpp_output.png - -# Load same image in browser tool -# Visually compare outputs -``` - -### GPU Timing - -Expected performance: -- 512×512: ~1-2 ms (integrated GPU) -- 1024×1024: ~3-5 ms -- 1920×1080: ~5-8 ms - -Slower than expected? Check: -- WebGPU enabled in browser -- Dedicated GPU selected (if available) -- No background tabs consuming GPU - ---- - -## Troubleshooting - -### "WebGPU not supported" - -- Update browser to latest version -- Enable WebGPU flag: `chrome://flags/#enable-unsafe-webgpu` -- Try Safari 18+ (native WebGPU on macOS) - -### "Invalid .bin file" - -- Check magic number: `hexdump -C weights.bin | head` -- Should start with: `43 4e 4e 32` ('CNN2') -- Re-export weights: `./training/export_cnn_v2_weights.py` - -### Black output / incorrect colors - -- Check blend slider (set to 1.0 for full CNN output) -- Verify training converged (loss < 0.01) -- Compare with C++ tool output - -### Shader compilation errors - -Open browser console (F12) for detailed errors. Common issues: -- Image too large (>4096×4096 not tested) -- Unsupported texture format (rare on modern GPUs) - ---- - -## Architecture - -**Pipeline:** -1. **Static Features Pass** - Generate 8D features (RGBD, UV, sin, bias) -2. **CNN Layer Passes** - Compute N layers with ping-pong textures -3. **Display Pass** - Unpack and render with view mode - -**Textures:** -- Input: RGBA8 (original image) -- Depth: R32F (uniform depth) -- Static features: RGBA32Uint (8×f16 packed) -- Layer buffers: RGBA32Uint (ping-pong) - -**Data-Driven Execution:** -- Layer count read from binary header -- Per-layer params (kernel size, channels, offsets) from binary -- Single CNN shader dispatched N times - ---- - -## Implemented Features - -**✓ Weights Metadata Panel:** -- Layer descriptions (kernel size, channels, weight count) -- Weight statistics (min/max per layer) -- File size and layer count - -**✓ Weights Visualization:** -- Per-layer kernel heatmaps (1px/weight) -- All input channels displayed horizontally -- Output channels stacked vertically -- Normalized grayscale display - -**✓ Layer Activation Viewer:** -- Static features (8D split into 0-3 and 4-7 views) -- All CNN layer outputs (Layer 0/1/2...) -- 4-channel split view (grayscale per channel) -- Mouse-driven 4× zoom view - -## TODO - -**Future Enhancements:** -- Weight distribution histograms per layer -- Activation statistics (min/max/mean overlay) -- Side-by-side diff mode (browser vs C++ output) -- Export rendered layers as PNG - ---- - -## Extensions (v2+) - -Planned enhancements: - -**Variable Feature Count:** -- Binary v2: Add `num_features` to header -- Shader: Dynamic feature array or multiple textures - -**Multi-Scale Input (Mip Levels):** -- Uncomment mip bindings in static shader -- No binary format change needed - -**8-bit Quantized Weights:** -- Binary version bump (format field already present) -- Add quantization codepath in `get_weight()` function -- 2× size reduction (~1.6 KB) - -**Pre-defined Test Images:** -- Dropdown menu with training/input/*.png -- Requires local file server - ---- - -## Size - -- HTML structure: ~2 KB -- CSS styling: ~2 KB -- JavaScript logic: ~10 KB (includes zoom + weights viz) -- Static shader: ~1 KB -- CNN shader: ~3 KB -- Display shader: ~1 KB -- Layer viz shader: ~2 KB -- Zoom shader: ~1 KB -- **Total: ~22 KB** (single file, no dependencies) - ---- - -## See Also - -- `doc/CNN_V2.md` - Architecture and design -- `doc/HOWTO.md` - Training workflows -- `training/export_cnn_v2_weights.py` - Binary format -- `src/effects/cnn_v2_effect.cc` - C++ reference implementation diff --git a/tools/cnn_v2_test/index.html b/tools/cnn_v2_test/index.html deleted file mode 100644 index e226d0c..0000000 --- a/tools/cnn_v2_test/index.html +++ /dev/null @@ -1,2049 +0,0 @@ -<!DOCTYPE html> -<html lang="en"> -<!-- - CNN v2 Testing Tool - WebGPU-based inference validator - - Architecture: - - Static features (8D): p0-p3 (parametric), uv_x, uv_y, sin(10*uv_x), bias (NOT a CNN layer) - - Layer 0: input RGBD (4D) + static (8D) = 12D → 4 channels - - Layer 1+: previous layer (4D) + static (8D) = 12D → 4 channels - - All CNN layers: uniform 12D input, 4D output (ping-pong buffer) - - Naming convention (matches train_cnn_v2.py / .wgsl / .cc): - - UI shows: "Static 0-3", "Static 4-7", "Layer 0", "Layer 1", "Layer 2" - - weights.layers[] array: Layer 0 = weights.layers[0], Layer 1 = weights.layers[1] - - Features: - - Input: PNG images or video files (MP4, WebM, etc.) - - Video playback: Play/Pause, frame-by-frame navigation (◄/► buttons) - - Video mode: Non-realtime processing (drops frames if CNN slower than playback) - - Side panel: .bin metadata display, weight statistics per layer - - Layer inspection: 4-channel grayscale split, intermediate layer visualization - - View modes: CNN output, original, diff (×10) - - Optimization: Layer viz updates only on pause/seek during video playback - - WGSL Shader Reuse: - - CNN_SHADER (inference), STATIC_SHADER, LAYER_VIZ_SHADER are inline for single-file deployment - - Can extract to .wgsl files for: better IDE support, testing, cross-tool reuse - - Tradeoff: extraction needs fetch() or build step, breaks single-file portability - - C++ sync: manual (WGSL ≠ GLSL) but logic identical ---> -<head> - <meta charset="UTF-8"> - <meta name="viewport" content="width=device-width, initial-scale=1.0"> - <title>CNN v2 Testing Tool</title> - <style> - * { margin: 0; padding: 0; box-sizing: border-box; } - body { - font-family: 'Courier New', monospace; - background: #1a1a1a; - color: #e0e0e0; - display: flex; - flex-direction: column; - height: 100vh; - overflow: hidden; - } - .header { - background: #2a2a2a; - padding: 16px; - border-bottom: 1px solid #404040; - display: flex; - align-items: center; - gap: 24px; - flex-wrap: wrap; - } - h1 { font-size: 18px; } - .controls { - display: flex; - gap: 16px; - align-items: center; - flex-wrap: wrap; - } - .control-group { - display: flex; - gap: 8px; - align-items: center; - } - .control-group label { font-size: 12px; } - input[type="range"] { width: 120px; } - input[type="number"] { width: 60px; background: #1a1a1a; color: #e0e0e0; border: 1px solid #404040; padding: 4px; } - .drop-zone { - border: 3px dashed #606060; - padding: 20px; - text-align: center; - cursor: pointer; - transition: all 0.2s; - font-size: 13px; - font-weight: bold; - background: #252525; - border-radius: 6px; - color: #4a9eff; - } - button { - background: #1a1a1a; - border: 1px solid #404040; - color: #e0e0e0; - padding: 6px 12px; - font-size: 12px; - font-family: 'Courier New', monospace; - cursor: pointer; - transition: all 0.2s; - border-radius: 4px; - } - button:hover { border-color: #606060; background: #252525; } - button:disabled { opacity: 0.3; cursor: not-allowed; } - video { display: none; } - .drop-zone:hover { border-color: #4a9eff; background: #2a3545; } - .drop-zone.active { border-color: #4a9eff; background: #1a2a3a; } - .drop-zone.error { border-color: #ff4a4a; background: #3a1a1a; } - .content { - flex: 1; - display: flex; - overflow: hidden; - gap: 1px; - background: #404040; - } - .left-sidebar { - width: 315px; - background: #2a2a2a; - overflow-y: auto; - display: flex; - flex-direction: column; - gap: 16px; - padding: 16px; - } - .main { - flex: 1; - display: flex; - justify-content: center; - align-items: center; - padding: 24px; - overflow: auto; - position: relative; - background: #1a1a1a; - } - .video-controls-float { - position: absolute; - top: 16px; - left: 50%; - transform: translateX(-50%); - display: flex; - gap: 8px; - background: rgba(42, 42, 42, 0.95); - padding: 8px 12px; - border-radius: 4px; - border: 1px solid #404040; - z-index: 100; - } - .bottom-controls-float { - position: absolute; - bottom: 16px; - left: 50%; - transform: translateX(-50%); - display: flex; - gap: 16px; - align-items: center; - background: rgba(42, 42, 42, 0.95); - padding: 8px 16px; - border-radius: 4px; - border: 1px solid #404040; - z-index: 100; - } - .bottom-controls-float .control-group { - display: flex; - gap: 8px; - align-items: center; - } - .bottom-controls-float #videoControls { - display: flex; - gap: 8px; - align-items: center; - padding-right: 16px; - border-right: 1px solid #404040; - } - .main.drop-active::after { - content: 'Drop PNG/video here'; - position: absolute; - inset: 24px; - display: flex; - align-items: center; - justify-content: center; - border: 3px dashed #4a9eff; - background: rgba(74, 158, 255, 0.1); - font-size: 24px; - color: #4a9eff; - pointer-events: none; - z-index: 10; - } - .sidebar { - width: 400px; - background: #2a2a2a; - overflow-y: auto; - display: flex; - flex-direction: column; - gap: 16px; - padding: 16px; - } - .panel { - border: 1px solid #404040; - border-radius: 4px; - overflow: hidden; - } - .panel.collapsed .panel-content { - display: none; - } - .panel-header { - background: #1a1a1a; - padding: 8px 12px; - font-size: 12px; - font-weight: bold; - border-bottom: 1px solid #404040; - } - .panel-content { - padding: 12px; - font-size: 11px; - } - .panel-content table { - width: 100%; - border-collapse: collapse; - } - .panel-content th { - text-align: left; - padding: 4px; - font-size: 10px; - color: #808080; - border-bottom: 1px solid #404040; - } - .panel-content td { - padding: 4px; - font-size: 10px; - } - .panel-content tr:hover { - background: #1a1a1a; - } - .layer-buttons { - display: flex; - flex-wrap: wrap; - gap: 6px; - margin-bottom: 12px; - } - .layer-buttons button { - background: #1a1a1a; - border: 1px solid #404040; - color: #e0e0e0; - padding: 6px 12px; - font-size: 10px; - font-family: 'Courier New', monospace; - cursor: pointer; - transition: all 0.2s; - } - .layer-buttons button:hover { - border-color: #606060; - background: #252525; - } - .layer-buttons button.active { - background: #4a9eff; - border-color: #4a9eff; - color: #1a1a1a; - } - .layer-buttons button:disabled { - opacity: 0.3; - cursor: not-allowed; - } - .layer-buttons button:disabled:hover { - border-color: #404040; - background: #1a1a1a; - } - .layer-grid { - display: grid; - grid-template-columns: repeat(4, 1fr); - gap: 4px; - margin-bottom: 12px; - } - .layer-view { - aspect-ratio: 1; - background: #1a1a1a; - border: 1px solid #404040; - display: flex; - flex-direction: column; - overflow: hidden; - } - .layer-preview { - background: #1a1a1a; - border: 1px solid #404040; - display: flex; - flex-direction: column; - overflow: hidden; - margin-top: 8px; - } - .layer-preview canvas { - width: 100%; - height: 100%; - image-rendering: pixelated; - } - .layer-view.active { - border: 2px solid #ffffff; - } - .layer-view canvas { - cursor: pointer; - } - .layer-view-label { - background: #2a2a2a; - padding: 4px; - font-size: 9px; - text-align: center; - border-bottom: 1px solid #404040; - } - .layer-view canvas { - width: 100%; - height: 100%; - image-rendering: pixelated; - } - canvas { - max-width: 100%; - max-height: 100%; - image-rendering: pixelated; - box-shadow: 0 4px 12px rgba(0,0,0,0.5); - } - .footer { - background: #2a2a2a; - border-top: 1px solid #404040; - font-size: 11px; - display: flex; - flex-direction: column; - gap: 8px; - } - .footer-top { - padding: 12px 16px 0; - display: flex; - justify-content: space-between; - } - .status { color: #4a9eff; } - .shortcuts { color: #808080; } - .console { - background: #1a1a1a; - padding: 8px 16px; - font-family: 'Courier New', monospace; - font-size: 10px; - color: #808080; - max-height: 100px; - overflow-y: auto; - border-top: 1px solid #404040; - } - .console-line { margin: 2px 0; } - .console-line.error { color: #ff4a4a; } - .console-line.info { color: #4a9eff; } - </style> -</head> -<body> - <div class="header"> - <h1>CNN v2 Testing Tool</h1> - </div> - <video id="videoSource" muted loop></video> - <div class="content"> - <div class="left-sidebar"> - <input type="file" id="weightsFile" accept=".bin" style="display: none;"> - <div class="drop-zone" id="weightsDrop" onclick="document.getElementById('weightsFile').click()"> - Drop .bin Weights or Click to Browse - </div> - <div class="panel" id="weightsInfoPanel"> - <div class="panel-header">Weights Info</div> - <div class="panel-content" id="weightsInfo"> - <p style="color: #808080; text-align: center;">No weights loaded</p> - </div> - </div> - <div class="panel" id="weightsVizPanel" style="display: none;"> - <div class="panel-header">Weights Visualization</div> - <div class="panel-content" id="weightsViz"> - <div class="layer-buttons" id="weightsLayerButtons"></div> - <canvas id="weightsCanvas" style="width: 100%; image-rendering: pixelated; border: 1px solid #404040;"></canvas> - </div> - </div> - <div class="panel"> - <div class="panel-content"> - <label for="mipLevel" style="font-size: 11px;">Mip Level:</label> - <select id="mipLevel" style="width: 100%; background: #1a1a1a; color: #e0e0e0; border: 1px solid #404040; padding: 4px; margin-top: 4px;"> - <option value="0">Mip 0 (original)</option> - <option value="1">Mip 1 (half res)</option> - <option value="2">Mip 2 (quarter res)</option> - </select> - </div> - </div> - </div> - <div class="main" id="mainDrop"> - <div class="bottom-controls-float"> - <div id="videoControls"> - <button id="playPauseBtn" disabled>Play</button> - <button id="stepBackBtn" disabled>◄ Frame</button> - <button id="stepForwardBtn" disabled>Frame ►</button> - </div> - <div class="control-group"> - <label>Blend:</label> - <input type="range" id="blend" min="0" max="1" step="0.01" value="1.0"> - <span id="blendValue">1.0</span> - </div> - <div class="control-group"> - <label>Depth:</label> - <input type="range" id="depth" min="0" max="1" step="0.01" value="1.0"> - <span id="depthValue">1.0</span> - </div> - <button id="savePngBtn">Save PNG</button> - </div> - <canvas id="canvas"></canvas> - </div> - <div class="sidebar"> - <div class="panel" style="flex: 1; display: flex; flex-direction: column; min-height: 0;"> - <div class="panel-header">Layer Visualization</div> - <div class="panel-content" id="layerViz" style="flex: 1; overflow: hidden;"> - <p style="color: #808080; text-align: center;">Load image + weights</p> - </div> - </div> - </div> - </div> - <div class="footer"> - <div class="footer-top"> - <span class="status" id="status">Drop PNG/video anywhere to begin</span> - <span class="shortcuts">[SPACE] Original | [D] Diff (×10)</span> - </div> - <div class="console" id="console"></div> - </div> - - <script> -// ============================================================================ -// EMBEDDED WEIGHTS & CONSTANTS -// ============================================================================ - -// Default pre-trained weights (base64-encoded binary format) -// Version 2: 4 layers (3×3, 5×5, 3×3, 3×3), 2496 f16 weights, mip_level=2 -const DEFAULT_WEIGHTS_B64 = 'Q05OMgIAAAAEAAAAwAkAAAIAAAADAAAADAAAAAQAAAAAAAAAsAEAAAUAAAAMAAAABAAAALABAACwBAAAAwAAAAwAAAAEAAAAYAYAALABAAADAAAADAAAAAQAAAAQCAAAsAEAAAU3faplMDmtR7gnMLqt6bSrLM4RCa/En4q257kVsmWz57aSHJMxz6wILJC0tLdBriWww7IULUehCClCo60dBiu1nWqsf60ZKn6ktCWKjrswATSfLwQunzJjKKWkN6hxLTMwbS2DJvgvUjFDL1YsQDFFL78ysC5OL/cvxC2kJ6qh0i1BLH2rzCrcKFUoeixTqwwopjD+rXmewCY6sYUtXCwwsaKqGjBcqoykKigRJYStaqjMp+siPi1BLI+tGatfK5Ii6C1qLY0tYSGFKz4wpzNdH1QuJDKmMJi0lLVAs0y2Q7YWtY21fLXusf+n8LDSsaethK3drB4rtSROKYOrLK53qrqu0REYLEUuVy1qEqohDSzgqk4sDKKSKi0clKcVKvupJ69rKTmw8q7qptatQK7OsFUw5Z5JKJ4udSp9LLQeui87LbcxljEgJ6Iw75jDLfUvIjCxnh0g763Lq/ItMqzDqP0sXCRcqnkl9qDlJUStSyR8oTuwA616IrAnNqo5JS4qDKeILmahyaHZI48tryiajuEs0aghLBcuny+aovQpAhj6Kqkwdy+8MZ0wLzBvKBStsrRAKJez+raaKAotBiVSqZqyk7b2sHO1e7cJsfGmQLACpWizBLP9LnWxYLWoJPeb/CY5ISokXqynJ4qtG6K1qpesL6zGqYssIDJRpnErRi3RL9kh1zBFLPkdGSNvKtEuvyywmgilbC43LNovbywCKj4pFzEbMmMuly2gMFYscCgzliIomSqZnpSnyK3hJJKsAasgJGMrfCyNqXwpqaYNq14wiyzWLrSn/yLbqm+tnauOpkKtRKdCrBcYQS0dnGAveqeBrD8sMiGpLkAugzEaLM6lLzAkL5YydzYnqGo15zh2MuSwJK0nqxI04jZ5LAs2TjilNeSc3yANLecrCzBCprUvfjUHMWCuFrAkItyq/an0JSUnvKnrrAosv5CRrTGvQKesntuur6v2rsyxzbCAsHYn1y5GrAGsASYUmawrpSLooRSy86sBqmaxAq67sD0lJalOKxOtkqx8H+wqgygMLhup8SzNKZuhcafWKUKs567KI1opDCsoplatAykJpc+skavUrK4p2iznLlMqcig4Le6mDKiaJpIsMiOgLGOtQqI7sFGworKfsTOq86ZIlru0dLCEoMqq4KzsI6I2MzixMocqSym8MwQtT7Njqrwy26rEthe2nTGxL/Gq+az8MPg1Tq6EqXmslqyArkKs/S73MqEwmyuzrUUxejLhKYaw0yUlMzgxAZULsZ4rhq8ssgarCjDTrPop0ywBLswwjbT7MMAxdq2fsEC04DZoOIovG7G4LwM1gTNnKDsuEbByrzyxvLLBKJgkGDQANSMy66wVrM21ebURriAluK5quFa3wLBsK2wvaDU7OEg3RDGWKVUzpTfPNG+tbrGcr3ytRKosr7yuCbB2rV6gZq3msWmtjqvmoNurP6YXrOIpf6l/J2irl6/iqK2jy6MCLkkhjSDQoAWWACo1JrWjP6nvKvmthay+KJ6rUqoKqaatHKyJrUOarydBo5yu/CUaKFoxFCW1CNgpri2WK02kgqvYqkotwqlIrdiiEa1aKZ2tXa6mrkax4KkYKp2vcKgErYsi2RvbqWapU6EAnMyqtyPBpYwdZyVZkwGl1yhhJ2QBPaUJqMmMJJ54IikpcqmUHzmacCDzq1Cr3yR9n8aizKlWKFiogapBFlknrimnHmemDqbVKHciNRyII5AsxZ0+Lf0Xmyh7LMIqDS2KK9EkxyxRHKgp2iL9K0QfxCwGLLEuwiqrLcWob6xpppasp6+lotypGrC9qdmpPKUuplagES2cpSyrsSyHJTMi3Kk4KWAlSCaqKNMtR626rKaoj6koI1wqeivGI9cpuqQ9KQUkZyEJKOmquyW0JymirSjhprWgkBpKLFykzZyloWSrNKxrGaCtMi1MqL6t56lLqu+wbbTetYkqYDR1rB0wqir/sWQwNas8N9E4wq+9I6WwT6xuMDy1yC9tM/Kwka+btK8vJisnIJWeUa30LRkwDaqIsNqzWK9lLnEzKjEMqYMuWy8uMs0qI6xKLjcvxicEqYCv06zrrLusKK/lMeMz8CyCMmqxO7AtNpW38zFzL5i2Wq19tkCuBaTlt8Kv85Mlsg6wWLfgstutzDJVNAqZxCywrQgspDYOMS0mGbQCuf63QS7GJ4GsBLizuRS0mKyiKKMkBbLXseCufCr4qKUpah7Vqh8tV6eqLLQoGy1bMNEu6i4fMD4wZSvbjwOpmCBzLMmeJKddoYqkIic6qpqRY6nNqDiwIq5dqcmndqbnKnGkSCjmKBUsriySrHWsZyTaG7smSKxAIwolIi2zLX6unK5KqXCwKq03qyarcKWMqQmmd6tIodWtH6UvLg2tTadPJOOp2iGgny0ufyy+L7AvNClhpiEpC6qMqqMp7KTopJ4mmB2ylM6mrKhfKiQrTyiiKdGoQqjKJ6Umxip/qDiq/ChgKtmqIiwOr+CunZF7Kfot36poqkcthCx+Ksapg5T5pn0oNqOPq4osMSbSqQQmGqgXKhEl3yV1piyswazLK7QoQBTaqU8lIS13Ldch+qQqJ2AsPKfmp3Ink5Z2HhosR5z4qLIoGqkNLCct2Ck3KPGnUC0oJBQq7agOKyaq0qsqpAap8SylLg4qriy6M3MqKCtdKpMjSi86KigsGCz/n2erEyu7J/QRVCkpILUwcC35LI8qxiw6Knoq5jAAKo8wnieqLF0vVTAYMZw4Jyx2t/ayTjGWMoGzKbwus1w4QRxeJse1dTGSNJGwmCrEJV8uQKygKe4gjSqkrLeydiaMroS0FrQms8Uygi28qe2uXS2Ko4q1d7ZxszEpiDSBMoc0STWpNc0xJKSvrMWm6bCKsOC3CrEOJNC1Ga5Qubi7U6/+NRQ0AqnSuFoySDmKtJS0b7KcNAMmqi45IbMvGzjeMg2qSioPKVWtSK6EpaA1UTckMt2m16nwM5E2oDHBsZ+pniVpMc4vQy1epXkqHifBl7Mu36T/KzQorix4JAOmWyqJFVUqq67doiot2CxYME8i2JxVKhQt5ioYJsWp1KiSpL0lhq1JpWAgbCweKW2o1CrCIMsrcghkHUqW3hiTI5osYqMlB+WaLy0uKNUooKx4qdEezqRlJEapyKuUoEmoZyT7nqcoo6v3n4yqZaGcpNElwij3IkinQiAFIFQK2ygqIoKsiZxEI6ukqCf7KFSkgqSTqjEq8JZLJPufXKmFkaEj36lCKj2qURxfKkQouaqQhRIrGSmepKin7Cl8KEcuKI+ip4Evz6xIF0woVK/yHLyfLSj0ny+oWywSJHWmQaEomWos6ZTMpPWlY61pqLelZqYGpAidcyzQE5kneBr1pnQkJSwIqWYpIabdKA8oHKroGeCnYplOKzAmC51LJ0emp6o+rXAofCkCKV4w4x1sKCYjrKAgKa0r+BcPJDMmP6o2JW4pIqqtm4srTqgHlLWlsBBepaqrKq27rBat9aTlot8qkaw2o5sl76ivKDkjNyjzKKWY5KlHrQCr8SjxquarXqrlKB2xyyfZL1Sqq7LWpxA04zZwMkyvUiyHMig1ay+GJqenVq1Ao1awVLHQnrEqxTD/LO8kKB+NH1grfKsPsY6u+aIELLaj4LBmLBU0wDOlM8ksdKjbqPSqQykHJmYodC+WMcYuSCJ7psYvNDTaLqWw/qy7Myw4xjTnMIouQTV9OJ81YSlbLiIx3TVuMUcokrDzI0ow8CQQr9IvDyxsLnk0OTVhLmmobLAULN4zkyyZsGC0LK01L3Upw52Jroywlix0MCwr5qkQJkot9aWzsYuui66HrHykMa9ZsDet96yBqXWvXbAXsraxIqgpsVOvtq5frF+iZa2WqROwcaP+qX2w+aW3rxWpI7Bwrlqu5K0LrxexX7DUrfOvhK3QrUGwP7BrsY2tU6yWr8qkpK18rn2rHCbloYmfaqM1nfSr7Sn1qjuk2KT2qyem4KXJJ4MdxaidqPWsa58zKTSsoKXAJUymz6rJpv+oGKsOJo2hSicHqA4oOiiRmr4k0BxBq8Ui16jTKvyq7ijmqHcpZanhHnGfMikxIiEk7S4Yq90sfKWSoZyntKg/qh+nJiifnAyvlKeXJMIdViKeoxEjLKvZpXymAqkhraCofK5SnTGmLqdkq7mjYCD8qV0qQKo0qrUo+KsZKVSs0iaULFUI8qS0mlWtiiqbGBegACwBoAErhaW1qMwqHSxfKVKpp6x7poiweKxCrdkivK48sJewrKdArHYnqyhoHbUnsagYK58qSjAgMcUwsCt0K/4rLC7mJGwtvStOMFQu0SzuJQUsBTBMLswqcJyEnVQsESn3ox2z9ai/qFqwES7tKP0vSChMoqQwVzR4LKaT+y/NK06q2y0LIi2wHrIcKZuzsrSHn/6xkrPssAovJzEipEQiDbDjr3SqIis5LGIoOSm6p1apeqGGrtAqJzCIJRuptqrApiktWTAwMB4xQizXKoIgASFFsLwweTHbLdQtqyzXoKYtay3SLeOke6wgoPWr/SpFKUEmDacWptSoMChJKm6s6azkHe+mfzFKKyamfi6bK/wr5atPqEMxUTAlKSeueiRxoSQjQqxQLRavgauKriOssymXLZOooa97pFoufTSppqgoVq05tEg196yCsQIy7bEitAItJ7RgtUEzxjGML/QmEKIlrPgjPDFaoTYoPDFcJRavtK4XrKmsk6zjsCwsTa4UsPQs9jI/I3ct1C6cMV+b5y7wJZ0tYTF9MGojdS/oLTShziM/MVmnxC8FKJUwRCUxIz8wiS4QLWipLCCYq9EseabMKnEll6kPqIawRq+xGcgjyCkgqKed7SB6qZcr6CwJLW+st6ePq7WuHycUrhqsSq7zsKuZtimgCXCrmKkqnIGp4LHNsX2wnqyBsH2xIbDhpwCzra1ss44wTCypKDCyyK23LRiwYKKPMJmxcaqZKcshCCYipoyxNa1Nsbwozi1+MB8lQ5mtsDel3jDnlbutxiPzsWmp5SpTHaqys7EstauTPqoRsOosf6g3sLOgeaAfKUIsWi/BJdosUSzdMM4pSy3kpGM0DjWvLWw0cjR4MWWqQaYMLo2rZSijJjstZiFaLBadMq0TseyjYi0VGsQt8yo5oZCgti/HMLciM6r3KgMk8K6OqKup9q0srT0xcaWMMMwra67qrhSfsZ3GrrIj2a2+pqSvdrEcrRQ0IDhgMB+PCDWVM8qjnJ5ZKOmw4C0dMGyuG6DGMQUvrq+Oq4UsTSzHMRg2ibbXs+Axa7N5sAqqnSoerQUmky8oKIiuUjGsoBitdKy9q6iw661pqg4thKnpkYmt+a3gseypGp5Co22fM6YSKJap66hwopmsmqhlrCMkZyiLL4KnGKupKvUmyCQbLFUrbSZerKahlaRoqCYm5SqYKW0rcS8WrAUkzaMcGlqpRK3bnresXy18IXapEKqHKFssXKCpKMUrfamapf4tKjBiKJGoU54HK+8q5qq4qVuiZiy4JuEsTixNMFQnlSSIIw4k1KzxpbMlDqyKqz6gra4SpcOw3a3Vq+qqC6tOq22eORvnpC8hRadkka2q/K7HHUiowawpqPInLyA0qYMlsihUqGGkWCb7K1WdWK5Dr5EhnKv5KHKlXqYnJ/2l9i0YKUYuMzHxpyCs/ChMkPEtwanxoFQqJi3Uq7Mseq3arXskWKc5pOAc7CZcqCwc5w7qKO4f3iaKIDsq/KRgLpWsQqn5rYYkxCWPoU0bx6hzGdkkqibtofEoxy8GpUupSCTiKiwvpij7LbiulqkErXetejFkL2+upqtUp0OwiLAPsdCpxLIlrKOyQ7C2r3utIg0drZEl2y6oLkquoaX4rCysAa9GDRCwKrHDsNivAbHsqtioqiGvrqgJE66Kqw4rzKyDKgaomp6TK2EsDyc0oOSol6NZJkmsvyxorMss5pR0KBquEixPpjsgXCpsnXQocq2MrfGmoivvLBeacahmLROpe6kcGCSfdC03qL6i6yitHHohrxzqq4UiP6JMqF8qThOshWAVUqHupDsoohQuJSkv/ywqLiwlNjG7o++hxi3vIKmleCdyrH6wYatdsPWsjLCNol+sSTDpryCptbBDK+qs4zBpLGc0Nqc1rdo09jX5MqsrHi2xKOad8igwJxAoeSsiqgkqdChcLOYxJzGlMkAsUzCuKzskTjAOKhuplqjHqf8wzDKYIGefNDISqd8pIC23Ltwu7zC9KgMsQDL/JcgrryYzLJ0oTSoyqpkmLax+KuejVyqxr08ulZ2XpyQr5yxRsEMpwzD0KmEqoihRC6mwF6xOplwmjSSmpMep0SvhpOEndCluqLyvtCGgo3unOyy9IXKtmZ9yIK8hlqohrEUtxh0XKH0sGi18p6coHa3Tow6psqa/JRUMU6yiKbUoXigQpo2i7C18q3ur6CnWrSateC3/KY+jlCJ6o6qr+x8VJUkSFadyAgGpji0xraytBSd+rYksTqDAHQAtxSjkqMAmNqxhqNesEi5uKsqlFqo9Kg6seizOrdusAasErjmtoKv8rb8ph6cYLnMmcKlCLJ6pjiuIKpkpKK1UKvyq3RhVpZac+izlrYitWB+DrI4omKOZKikiZS1Fqicf+q25rJmsqKrYrNGt0JWRLWel2KfLqQ=='; - -// Reusable fullscreen quad vertex shader (2 triangles covering NDC) -const FULLSCREEN_QUAD_VS = ` -@vertex -fn vs_main(@builtin(vertex_index) idx: u32) -> @builtin(position) vec4<f32> { - var pos = array<vec2<f32>, 6>( - vec2<f32>(-1.0, -1.0), vec2<f32>(1.0, -1.0), vec2<f32>(-1.0, 1.0), - vec2<f32>(-1.0, 1.0), vec2<f32>(1.0, -1.0), vec2<f32>(1.0, 1.0) - ); - return vec4<f32>(pos[idx], 0.0, 1.0); -}`; - -// ============================================================================ -// WGSL SHADERS -// ============================================================================ - -// Static features: 7D parametric features (RGBD + UV + sin(10*uv_x) + bias) -const STATIC_SHADER = ` -@group(0) @binding(0) var input_tex: texture_2d<f32>; -@group(0) @binding(1) var linear_sampler: sampler; -@group(0) @binding(2) var depth_tex: texture_2d<f32>; -@group(0) @binding(3) var output_tex: texture_storage_2d<rgba32uint, write>; -@group(0) @binding(4) var<uniform> mip_level: u32; - -@compute @workgroup_size(8, 8) -fn main(@builtin(global_invocation_id) id: vec3<u32>) { - let coord = vec2<i32>(id.xy); - let dims = textureDimensions(input_tex); - if (coord.x >= i32(dims.x) || coord.y >= i32(dims.y)) { return; } - - // Use normalized UV coords with linear sampler (bilinear filtering) - let uv = (vec2<f32>(coord) + 0.5) / vec2<f32>(dims); - let rgba = textureSampleLevel(input_tex, linear_sampler, uv, f32(mip_level)); - - let p0 = rgba.r; - let p1 = rgba.g; - let p2 = rgba.b; - let p3 = textureLoad(depth_tex, coord, 0).r; - - let uv_x = f32(coord.x) / f32(dims.x); - let uv_y = f32(coord.y) / f32(dims.y); - let sin20_y = sin(20.0 * uv_y); - let bias = 1.0; - - let packed = vec4<u32>( - pack2x16float(vec2<f32>(p0, p1)), - pack2x16float(vec2<f32>(p2, p3)), - pack2x16float(vec2<f32>(uv_x, uv_y)), - pack2x16float(vec2<f32>(sin20_y, bias)) - ); - textureStore(output_tex, coord, packed); -}`; - -const CNN_SHADER = ` -struct LayerParams { - kernel_size: u32, - in_channels: u32, - out_channels: u32, - weight_offset: u32, - is_output_layer: u32, - blend_amount: f32, - is_layer_0: u32, -} - -@group(0) @binding(0) var static_features: texture_2d<u32>; -@group(0) @binding(1) var layer_input: texture_2d<u32>; -@group(0) @binding(2) var output_tex: texture_storage_2d<rgba32uint, write>; -@group(0) @binding(3) var<storage, read> weights_buffer: array<u32>; -@group(0) @binding(4) var<uniform> params: LayerParams; -@group(0) @binding(5) var original_input: texture_2d<f32>; - -fn unpack_static_features(coord: vec2<i32>) -> array<f32, 8> { - let packed = textureLoad(static_features, coord, 0); - let v0 = unpack2x16float(packed.x); - let v1 = unpack2x16float(packed.y); - let v2 = unpack2x16float(packed.z); - let v3 = unpack2x16float(packed.w); - return array<f32, 8>(v0.x, v0.y, v1.x, v1.y, v2.x, v2.y, v3.x, v3.y); -} - -fn unpack_layer_channels(coord: vec2<i32>) -> vec4<f32> { - let packed = textureLoad(layer_input, coord, 0); - let v0 = unpack2x16float(packed.x); - let v1 = unpack2x16float(packed.y); - return vec4<f32>(v0.x, v0.y, v1.x, v1.y); -} - -fn pack_channels(values: vec4<f32>) -> vec4<u32> { - return vec4<u32>( - pack2x16float(vec2<f32>(values.x, values.y)), - pack2x16float(vec2<f32>(values.z, values.w)), - 0u, - 0u - ); -} - -fn get_weight(idx: u32) -> f32 { - let pair_idx = idx / 2u; - let packed = weights_buffer[pair_idx]; - let unpacked = unpack2x16float(packed); - return select(unpacked.y, unpacked.x, (idx & 1u) == 0u); -} - -@compute @workgroup_size(8, 8) -fn main(@builtin(global_invocation_id) id: vec3<u32>) { - let coord = vec2<i32>(id.xy); - let dims = textureDimensions(static_features); - if (coord.x >= i32(dims.x) || coord.y >= i32(dims.y)) { return; } - - let kernel_size = params.kernel_size; - let in_channels = params.in_channels; // Always 12 (4 prev + 8 static) - let out_channels = params.out_channels; // Always 4 - let weight_offset = params.weight_offset; - let is_output = params.is_output_layer != 0u; - let kernel_radius = i32(kernel_size / 2u); - - let static_feat = unpack_static_features(coord); - - var output: vec4<f32> = vec4<f32>(0.0); - for (var c: u32 = 0u; c < 4u; c++) { - var sum: f32 = 0.0; - for (var ky: i32 = -kernel_radius; ky <= kernel_radius; ky++) { - for (var kx: i32 = -kernel_radius; kx <= kernel_radius; kx++) { - let sample_coord = coord + vec2<i32>(kx, ky); - let clamped = vec2<i32>( - clamp(sample_coord.x, 0, i32(dims.x) - 1), - clamp(sample_coord.y, 0, i32(dims.y) - 1) - ); - let static_local = unpack_static_features(clamped); - let layer_local = unpack_layer_channels(clamped); - - let ky_idx = u32(ky + kernel_radius); - let kx_idx = u32(kx + kernel_radius); - let spatial_idx = ky_idx * kernel_size + kx_idx; - - // Previous layer channels (4D) - for (var i: u32 = 0u; i < 4u; i++) { - let w_idx = weight_offset + - c * in_channels * kernel_size * kernel_size + - i * kernel_size * kernel_size + spatial_idx; - sum += get_weight(w_idx) * layer_local[i]; - } - - // Static features (8D) - for (var i: u32 = 0u; i < 8u; i++) { - let w_idx = weight_offset + - c * in_channels * kernel_size * kernel_size + - (4u + i) * kernel_size * kernel_size + spatial_idx; - sum += get_weight(w_idx) * static_local[i]; - } - } - } - - if (is_output || params.is_layer_0 != 0u) { - output[c] = 1.0 / (1.0 + exp(-sum)); // Sigmoid [0,1] - } else { - output[c] = max(0.0, sum); // ReLU - } - } - - if (is_output) { - let original = textureLoad(original_input, coord, 0).rgb; - let result_rgb = vec3<f32>(output.x, output.y, output.z); - let blended = mix(original, result_rgb, params.blend_amount); - output.x = blended.r; - output.y = blended.g; - output.z = blended.b; - } - - textureStore(output_tex, coord, pack_channels(output)); -}`; - -const DISPLAY_SHADER = ` -@group(0) @binding(0) var result_tex: texture_2d<u32>; -@group(0) @binding(1) var original_tex: texture_2d<f32>; -@group(0) @binding(2) var<uniform> mode: u32; - -@vertex -fn vs_main(@builtin(vertex_index) idx: u32) -> @builtin(position) vec4<f32> { - var pos = array<vec2<f32>, 6>( - vec2<f32>(-1.0, -1.0), vec2<f32>(1.0, -1.0), vec2<f32>(-1.0, 1.0), - vec2<f32>(-1.0, 1.0), vec2<f32>(1.0, -1.0), vec2<f32>(1.0, 1.0) - ); - return vec4<f32>(pos[idx], 0.0, 1.0); -} - -@fragment -fn fs_main(@builtin(position) pos: vec4<f32>) -> @location(0) vec4<f32> { - let coord = vec2<i32>(pos.xy); - let packed = textureLoad(result_tex, coord, 0); - let v0 = unpack2x16float(packed.x); - let v1 = unpack2x16float(packed.y); - let result = vec3<f32>(v0.x, v0.y, v1.x); - - if (mode == 0u) { - return vec4<f32>(result, 1.0); - } else if (mode == 1u) { - let original = textureLoad(original_tex, coord, 0).rgb; - return vec4<f32>(original, 1.0); - } else { - let original = textureLoad(original_tex, coord, 0).rgb; - let diff = abs(result - original) * 10.0; - return vec4<f32>(diff, 1.0); - } -}`; - -const LAYER_VIZ_SHADER = ` -@group(0) @binding(0) var layer_tex: texture_2d<u32>; -@group(0) @binding(1) var<uniform> viz_params: vec2<f32>; // x=channel_idx, y=scale - -@vertex -fn vs_main(@builtin(vertex_index) idx: u32) -> @builtin(position) vec4<f32> { - var pos = array<vec2<f32>, 6>( - vec2<f32>(-1.0, -1.0), vec2<f32>(1.0, -1.0), vec2<f32>(-1.0, 1.0), - vec2<f32>(-1.0, 1.0), vec2<f32>(1.0, -1.0), vec2<f32>(1.0, 1.0) - ); - return vec4<f32>(pos[idx], 0.0, 1.0); -} - -@fragment -fn fs_main(@builtin(position) pos: vec4<f32>) -> @location(0) vec4<f32> { - let coord = vec2<i32>(pos.xy); - let dims = textureDimensions(layer_tex); - - let channel = u32(viz_params.x); - - // DEBUG MODE 1: Texture coordinates (channel 10) - if (channel == 10u) { - let uv = vec2<f32>(f32(coord.x) / f32(dims.x), f32(coord.y) / f32(dims.y)); - return vec4<f32>(uv.x, uv.y, 0.0, 1.0); - } - - let packed = textureLoad(layer_tex, coord, 0); - - // DEBUG MODE 2: Raw packed data (channel 11) - if (channel == 11u) { - let raw_val = f32(packed.x) / 4294967295.0; - return vec4<f32>(raw_val, raw_val, raw_val, 1.0); - } - - let v0 = unpack2x16float(packed.x); - let v1 = unpack2x16float(packed.y); - let v2 = unpack2x16float(packed.z); - let v3 = unpack2x16float(packed.w); - - // DEBUG MODE 3: First unpacked value (channel 12) - if (channel == 12u) { - return vec4<f32>(v0.x, v0.x, v0.x, 1.0); - } - - var channels: array<f32, 8>; - channels[0] = v0.x; - channels[1] = v0.y; - channels[2] = v1.x; - channels[3] = v1.y; - channels[4] = v2.x; - channels[5] = v2.y; - channels[6] = v3.x; - channels[7] = v3.y; - - let scale = viz_params.y; - - let idx = min(channel, 7u); - let raw = channels[idx]; - - // Apply scale: multiply and clamp to [0, 1] - let val = clamp(raw * scale, 0.0, 1.0); - - return vec4<f32>(val, val, val, 1.0); -}`; - -class CNNTester { - constructor() { - this.canvas = document.getElementById('canvas'); - this.status = document.getElementById('status'); - this.console = document.getElementById('console'); - this.image = null; - this.video = document.getElementById('videoSource'); - this.weights = null; - this.viewMode = 0; - this.blendAmount = 1.0; - this.depth = 1.0; - this.currentLayerIdx = null; - this.currentChannelOffset = null; - this.isVideo = false; - this.fps = 30; - this.isProcessing = false; - this.mipLevel = 0; - this.selectedChannel = 0; - this.init(); - } - - log(msg, type = 'info') { - const line = document.createElement('div'); - line.className = `console-line ${type}`; - line.textContent = `[${new Date().toLocaleTimeString()}] ${msg}`; - this.console.appendChild(line); - this.console.scrollTop = this.console.scrollHeight; - } - - async init() { - if (!navigator.gpu) { - this.setStatus('WebGPU not supported', true); - this.log('WebGPU not supported in this browser', 'error'); - return; - } - - try { - this.adapter = await navigator.gpu.requestAdapter(); - this.device = await this.adapter.requestDevice(); - this.context = this.canvas.getContext('webgpu'); - this.format = navigator.gpu.getPreferredCanvasFormat(); - this.log('WebGPU initialized successfully'); - } catch (e) { - this.setStatus(`GPU init failed: ${e.message}`, true); - this.log(`GPU initialization failed: ${e.message}`, 'error'); - } - } - - setStatus(msg, isError = false) { - this.status.textContent = msg; - this.status.style.color = isError ? '#ff4a4a' : '#4a9eff'; - } - - // Get current source dimensions (video or image) - getDimensions() { - if (this.isVideo) { - return { width: this.video.videoWidth, height: this.video.videoHeight }; - } - return { width: this.image.width, height: this.image.height }; - } - - // Enable/disable video playback controls - setVideoControlsEnabled(enabled) { - ['playPauseBtn', 'stepBackBtn', 'stepForwardBtn'].forEach(id => - document.getElementById(id).disabled = !enabled - ); - } - - parseWeights(buffer) { - const view = new DataView(buffer); - const magic = view.getUint32(0, true); - if (magic !== 0x32_4E_4E_43) { - throw new Error('Invalid .bin file (bad magic)'); - } - - const version = view.getUint32(4, true); - const numLayers = view.getUint32(8, true); - const totalWeights = view.getUint32(12, true); - - // Version 2: added mip_level field (20-byte header) - let mipLevel = 0; - let headerSize = 16; - if (version === 2) { - mipLevel = view.getUint32(16, true); - headerSize = 20; - this.log(`Binary header: version=${version}, layers=${numLayers}, weights=${totalWeights}, mip_level=${mipLevel}`); - } else if (version === 1) { - this.log(`Binary header: version=${version}, layers=${numLayers}, weights=${totalWeights}`); - } else { - throw new Error(`Unsupported binary version: ${version}`); - } - - const layers = []; - for (let i = 0; i < numLayers; i++) { - const offset = headerSize + i * 20; - const layer = { - kernelSize: view.getUint32(offset, true), - inChannels: view.getUint32(offset + 4, true), - outChannels: view.getUint32(offset + 8, true), - weightOffset: view.getUint32(offset + 12, true), - weightCount: view.getUint32(offset + 16, true), - }; - layers.push(layer); - this.log(` Layer ${i}: ${layer.inChannels}→${layer.outChannels}, kernel=${layer.kernelSize}×${layer.kernelSize}, weights=${layer.weightCount}`); - } - - const weightsOffset = headerSize + numLayers * 20; - const weights = new Uint32Array(buffer.slice(weightsOffset)); - - // Calculate min/max per layer - for (let i = 0; i < numLayers; i++) { - const layer = layers[i]; - let min = Infinity, max = -Infinity; - const startIdx = layer.weightOffset; - const endIdx = startIdx + layer.weightCount; - - for (let j = startIdx; j < endIdx; j++) { - const pairIdx = Math.floor(j / 2); - const packed = weights[pairIdx]; - const unpacked = this.unpackF16(packed); - const val = (j % 2 === 0) ? unpacked[0] : unpacked[1]; - min = Math.min(min, val); - max = Math.max(max, val); - } - - layer.min = min; - layer.max = max; - this.log(` Layer ${i} range: [${min.toFixed(4)}, ${max.toFixed(4)}]`); - } - - let nonZero = 0; - for (let i = 0; i < weights.length; i++) { - if (weights[i] !== 0) nonZero++; - } - this.log(` Weight buffer: ${weights.length} u32 (${nonZero} non-zero)`); - - return { version, layers, weights, mipLevel, fileSize: buffer.byteLength }; - } - - unpackF16(packed) { - const lo = packed & 0xFFFF; - const hi = (packed >> 16) & 0xFFFF; - const toFloat = (bits) => { - const sign = (bits >> 15) & 1; - const exp = (bits >> 10) & 0x1F; - const frac = bits & 0x3FF; - if (exp === 0) return (sign ? -1 : 1) * Math.pow(2, -14) * (frac / 1024); - if (exp === 31) return frac ? NaN : (sign ? -Infinity : Infinity); - return (sign ? -1 : 1) * Math.pow(2, exp - 15) * (1 + frac / 1024); - }; - return [toFloat(lo), toFloat(hi)]; - } - - async loadImage(file) { - const img = await createImageBitmap(file); - this.image = img; - this.isVideo = false; - this.canvas.width = img.width; - this.canvas.height = img.height; - this.setVideoControlsEnabled(false); - this.log(`Loaded image: ${file.name} (${img.width}×${img.height})`); - if (this.weights) { - this.setStatus(`Ready: ${img.width}×${img.height}`); - this.run(); - } else { - this.setStatus(`Image loaded (${img.width}×${img.height}) - drop .bin weights to process`); - this.displayOriginal(); - } - } - - // Video loading: wait for metadata, then first frame decode (readyState≥2) - async loadVideo(file) { - return new Promise((resolve, reject) => { - this.video.src = URL.createObjectURL(file); - - this.video.onloadedmetadata = () => { - const w = this.video.videoWidth; - const h = this.video.videoHeight; - if (w === 0 || h === 0) { - reject(new Error('Video has invalid dimensions')); - return; - } - - this.isVideo = true; - this.canvas.width = w; - this.canvas.height = h; - this.fps = 30; - this.log(`Loaded video: ${file.name} (${w}×${h}, ${this.video.duration.toFixed(1)}s)`); - this.setVideoControlsEnabled(true); - - // Set up event handlers - this.video.onpause = () => { document.getElementById('playPauseBtn').textContent = 'Play'; }; - this.video.onplay = () => { document.getElementById('playPauseBtn').textContent = 'Pause'; this.playbackLoop(); }; - - // Wait for first frame to be decoded before displaying - const displayFirstFrame = () => { - this.video.onseeked = () => { if (!this.isProcessing) this.processVideoFrame(); }; - if (this.video.readyState >= 2) { // HAVE_CURRENT_DATA or better - if (this.weights) { - this.setStatus(`Ready: ${w}×${h}`); - this.processVideoFrame().then(() => resolve()); - } else { - this.setStatus(`Video loaded - drop .bin weights to process`); - this.displayOriginal(); - resolve(); - } - } else { - setTimeout(displayFirstFrame, 50); // Poll until frame ready - } - }; - - this.video.onseeked = displayFirstFrame; - this.video.currentTime = 0; - }; - - this.video.onerror = () => reject(new Error('Failed to load video')); - }); - } - - // Video playback loop (non-realtime, drops frames if CNN slow) - playbackLoop() { - if (this.video.paused || this.video.ended) return; - if (!this.isProcessing) this.processVideoFrame(); - requestAnimationFrame(() => this.playbackLoop()); - } - - // Process current video frame through CNN pipeline - async processVideoFrame() { - if (!this.weights || this.isProcessing) return; - this.isProcessing = true; - await this.run(); - this.isProcessing = false; - } - - // Video controls - togglePlayPause() { - this.video.paused ? this.video.play() : this.video.pause(); - } - - stepFrame(direction) { - if (!this.isVideo) return; - this.video.pause(); - this.video.currentTime = Math.max(0, Math.min(this.video.duration, - this.video.currentTime + direction / this.fps)); - } - - async loadWeights(file) { - const buffer = await file.arrayBuffer(); - this.weights = this.parseWeights(buffer); - this.weightsBuffer = buffer; - this.mipLevel = this.weights.mipLevel; // Set mip level from binary format - this.log(`Loaded weights: ${file.name} (${this.weights.layers.length} layers, ${(buffer.byteLength/1024).toFixed(1)} KB)`); - - // Update UI dropdown to reflect loaded mip level - const mipLevelSelect = document.getElementById('mipLevel'); - if (mipLevelSelect) { - mipLevelSelect.value = this.mipLevel.toString(); - } - - this.updateWeightsPanel(); - if (this.image) { - this.setStatus(`Ready: ${this.image.width}×${this.image.height}`); - this.run(); - } else { - this.setStatus('Weights loaded - drop PNG image to process'); - } - } - - updateWeightsPanel() { - const panel = document.getElementById('weightsInfo'); - const { version, layers, mipLevel, fileSize } = this.weights; - - let html = ` - <div style="margin-bottom: 12px;"> - <div><strong>File Size:</strong> ${(fileSize / 1024).toFixed(2)} KB</div> - <div><strong>Version:</strong> ${version}</div> - <div><strong>CNN Layers:</strong> ${layers.length}</div> - <div><strong>Mip Level:</strong> ${mipLevel} (p0-p3 features)</div> - <div style="font-size: 9px; color: #808080; margin-top: 4px;">Static features (input) + ${layers.length} conv layers</div> - </div> - <table> - <thead> - <tr> - <th>Layer</th> - <th>Size</th> - <th>Weights</th> - <th>Min</th> - <th>Max</th> - </tr> - </thead> - <tbody> - `; - - // Display layers as "Layer 0", "Layer 1", etc. (matching codebase convention) - for (let i = 0; i < layers.length; i++) { - const l = layers[i]; - html += ` - <tr> - <td>Layer ${i}</td> - <td>${l.inChannels}→${l.outChannels} (${l.kernelSize}×${l.kernelSize})</td> - <td>${l.weightCount}</td> - <td>${l.min.toFixed(3)}</td> - <td>${l.max.toFixed(3)}</td> - </tr> - `; - } - - html += ` - </tbody> - </table> - `; - - panel.innerHTML = html; - - // Show weights visualization panel and create layer buttons - const weightsVizPanel = document.getElementById('weightsVizPanel'); - weightsVizPanel.style.display = 'block'; - - const weightsLayerButtons = document.getElementById('weightsLayerButtons'); - let buttonsHtml = ''; - for (let i = 0; i < layers.length; i++) { - buttonsHtml += `<button onclick="tester.visualizeWeights(${i})" id="weightsBtn${i}">Layer ${i}</button>`; - } - weightsLayerButtons.innerHTML = buttonsHtml; - - // Auto-select first layer - this.visualizeWeights(0); - } - - generateMipmaps(texture, width, height) { - if (!this.mipmapPipeline) { - const mipmapShader = FULLSCREEN_QUAD_VS + ` - @group(0) @binding(0) var src: texture_2d<f32>; - @fragment - fn fs_main(@builtin(position) pos: vec4<f32>) -> @location(0) vec4<f32> { - let coord = vec2<i32>(i32(pos.x) * 2, i32(pos.y) * 2); - var sum = vec4<f32>(0.0); - for (var y: i32 = 0; y < 2; y++) { - for (var x: i32 = 0; x < 2; x++) { - sum += textureLoad(src, coord + vec2<i32>(x, y), 0); - } - } - return sum * 0.25; - } - `; - this.mipmapPipeline = this.device.createRenderPipeline({ - layout: 'auto', - vertex: { module: this.device.createShaderModule({ code: mipmapShader }), entryPoint: 'vs_main' }, - fragment: { - module: this.device.createShaderModule({ code: mipmapShader }), - entryPoint: 'fs_main', - targets: [{ format: 'rgba8unorm' }] - } - }); - } - - const encoder = this.device.createCommandEncoder(); - - for (let mip = 1; mip < 3; mip++) { - const mipWidth = Math.max(1, width >> mip); - const mipHeight = Math.max(1, height >> mip); - - const bindGroup = this.device.createBindGroup({ - layout: this.mipmapPipeline.getBindGroupLayout(0), - entries: [ - { binding: 0, resource: texture.createView({ baseMipLevel: mip - 1, mipLevelCount: 1 }) } - ] - }); - - const renderPass = encoder.beginRenderPass({ - colorAttachments: [{ - view: texture.createView({ baseMipLevel: mip, mipLevelCount: 1 }), - loadOp: 'clear', - storeOp: 'store' - }] - }); - - renderPass.setPipeline(this.mipmapPipeline); - renderPass.setBindGroup(0, bindGroup); - renderPass.setViewport(0, 0, mipWidth, mipHeight, 0, 1); - renderPass.draw(6); - renderPass.end(); - } - - this.device.queue.submit([encoder.finish()]); - } - - displayOriginal() { - const source = this.isVideo ? this.video : this.image; - if (!source || !this.device) return; - - const { width, height } = this.getDimensions(); - this.context.configure({ device: this.device, format: this.format }); - - const inputTex = this.device.createTexture({ - size: [width, height], - format: 'rgba8unorm', - usage: GPUTextureUsage.TEXTURE_BINDING | GPUTextureUsage.COPY_DST | GPUTextureUsage.RENDER_ATTACHMENT - }); - - this.device.queue.copyExternalImageToTexture( - { source: source }, - { texture: inputTex }, - [width, height] - ); - - const simpleShader = FULLSCREEN_QUAD_VS + ` - @group(0) @binding(0) var tex: texture_2d<f32>; - @fragment - fn fs_main(@builtin(position) pos: vec4<f32>) -> @location(0) vec4<f32> { - let coord = vec2<i32>(pos.xy); - return textureLoad(tex, coord, 0); - } - `; - - const pipeline = this.device.createRenderPipeline({ - layout: 'auto', - vertex: { module: this.device.createShaderModule({ code: simpleShader }), entryPoint: 'vs_main' }, - fragment: { - module: this.device.createShaderModule({ code: simpleShader }), - entryPoint: 'fs_main', - targets: [{ format: this.format }] - } - }); - - const bindGroup = this.device.createBindGroup({ - layout: pipeline.getBindGroupLayout(0), - entries: [{ binding: 0, resource: inputTex.createView() }] - }); - - const encoder = this.device.createCommandEncoder(); - const renderPass = encoder.beginRenderPass({ - colorAttachments: [{ - view: this.context.getCurrentTexture().createView(), - loadOp: 'clear', - storeOp: 'store' - }] - }); - renderPass.setPipeline(pipeline); - renderPass.setBindGroup(0, bindGroup); - renderPass.draw(6); - renderPass.end(); - - this.device.queue.submit([encoder.finish()]); - } - - // Run CNN inference pipeline on current source (image or video frame) - async run() { - const t0 = performance.now(); - const source = this.isVideo ? this.video : this.image; - if (!source) return; - const { width, height } = this.getDimensions(); - - this.context.configure({ device: this.device, format: this.format }); - - // Create persistent input texture for original view with mipmaps - if (this.inputTexture) this.inputTexture.destroy(); - this.inputTexture = this.device.createTexture({ - size: [width, height], - format: 'rgba8unorm', - mipLevelCount: 3, - usage: GPUTextureUsage.TEXTURE_BINDING | GPUTextureUsage.COPY_DST | GPUTextureUsage.RENDER_ATTACHMENT - }); - - this.device.queue.copyExternalImageToTexture( - { source: source }, - { texture: this.inputTexture, mipLevel: 0 }, - [width, height] - ); - - // Generate mipmaps - this.generateMipmaps(this.inputTexture, width, height); - - const staticTex = this.device.createTexture({ - size: [width, height], - format: 'rgba32uint', - usage: GPUTextureUsage.STORAGE_BINDING | GPUTextureUsage.TEXTURE_BINDING | GPUTextureUsage.COPY_SRC - }); - - // Create one texture per layer output (static + all CNN layers) - this.layerOutputs = []; - const numLayers = this.weights.layers.length + 1; // +1 for static features - const layerTextures = []; - for (let i = 0; i < numLayers; i++) { - layerTextures.push(this.device.createTexture({ - size: [width, height], - format: 'rgba32uint', - usage: GPUTextureUsage.STORAGE_BINDING | GPUTextureUsage.TEXTURE_BINDING | GPUTextureUsage.COPY_DST - })); - } - - // Ping-pong buffers for computation - const computeTextures = [ - this.device.createTexture({ - size: [width, height], - format: 'rgba32uint', - usage: GPUTextureUsage.STORAGE_BINDING | GPUTextureUsage.TEXTURE_BINDING | GPUTextureUsage.COPY_SRC - }), - this.device.createTexture({ - size: [width, height], - format: 'rgba32uint', - usage: GPUTextureUsage.STORAGE_BINDING | GPUTextureUsage.TEXTURE_BINDING | GPUTextureUsage.COPY_SRC - }) - ]; - - const weightsGPU = this.device.createBuffer({ - size: this.weightsBuffer.byteLength, - usage: GPUBufferUsage.STORAGE | GPUBufferUsage.COPY_DST - }); - this.device.queue.writeBuffer(weightsGPU, 0, this.weightsBuffer); - const staticPipeline = this.device.createComputePipeline({ - layout: 'auto', - compute: { module: this.device.createShaderModule({ code: STATIC_SHADER }), entryPoint: 'main' } - }); - - const cnnPipeline = this.device.createComputePipeline({ - layout: 'auto', - compute: { module: this.device.createShaderModule({ code: CNN_SHADER }), entryPoint: 'main' } - }); - - const displayPipeline = this.device.createRenderPipeline({ - layout: 'auto', - vertex: { module: this.device.createShaderModule({ code: DISPLAY_SHADER }), entryPoint: 'vs_main' }, - fragment: { - module: this.device.createShaderModule({ code: DISPLAY_SHADER }), - entryPoint: 'fs_main', - targets: [{ format: this.format }] - } - }); - - const encoder = this.device.createCommandEncoder(); - - const mipLevelBuffer = this.device.createBuffer({ - size: 4, - usage: GPUBufferUsage.UNIFORM | GPUBufferUsage.COPY_DST - }); - this.device.queue.writeBuffer(mipLevelBuffer, 0, new Uint32Array([this.mipLevel])); - - if (!this.pointSampler) { - this.pointSampler = this.device.createSampler({ - magFilter: 'linear', - minFilter: 'linear', - mipmapFilter: 'linear' - }); - } - - // Extract depth from alpha channel (or 1.0 if no alpha) - const depthTex = this.device.createTexture({ - size: [width, height, 1], - format: 'r32float', - usage: GPUTextureUsage.TEXTURE_BINDING | GPUTextureUsage.COPY_DST - }); - - // Read image data to extract alpha channel - const tempCanvas = document.createElement('canvas'); - tempCanvas.width = width; - tempCanvas.height = height; - const tempCtx = tempCanvas.getContext('2d'); - tempCtx.drawImage(source, 0, 0, width, height); - const imageData = tempCtx.getImageData(0, 0, width, height); - const pixels = imageData.data; - - // Extract alpha channel (RGBA format: every 4th byte) - const depthData = new Float32Array(width * height); - for (let i = 0; i < width * height; i++) { - depthData[i] = pixels[i * 4 + 3] / 255.0; // Alpha channel [0, 255] → [0, 1] - } - - this.device.queue.writeTexture( - { texture: depthTex }, - depthData, - { bytesPerRow: width * 4 }, - [width, height, 1] - ); - - const staticBG = this.device.createBindGroup({ - layout: staticPipeline.getBindGroupLayout(0), - entries: [ - { binding: 0, resource: this.inputTexture.createView() }, - { binding: 1, resource: this.pointSampler }, - { binding: 2, resource: depthTex.createView() }, // Depth from alpha (matches training) - { binding: 3, resource: staticTex.createView() }, - { binding: 4, resource: { buffer: mipLevelBuffer } } - ] - }); - - const staticPass = encoder.beginComputePass(); - staticPass.setPipeline(staticPipeline); - staticPass.setBindGroup(0, staticBG); - staticPass.dispatchWorkgroups(Math.ceil(width / 8), Math.ceil(height / 8)); - staticPass.end(); - - // Copy static features to persistent storage (visualization index 0, shown as Static 0-3 / Static 4-7) - encoder.copyTextureToTexture( - { texture: staticTex }, - { texture: layerTextures[0] }, - [width, height] - ); - this.layerOutputs.push(layerTextures[0]); - - let srcTex = staticTex; - let dstTex = computeTextures[0]; - - for (let i = 0; i < this.weights.layers.length; i++) { - const layer = this.weights.layers[i]; - const isOutput = i === this.weights.layers.length - 1; - - // Calculate absolute weight offset in f16 units (add header offset) - // Version 1: 4 u32 header, Version 2: 5 u32 header - const headerSizeU32 = (this.weights.version === 1) ? 4 : 5; - const headerOffsetU32 = headerSizeU32 + this.weights.layers.length * 5; // Header + layer info in u32 - const absoluteWeightOffset = headerOffsetU32 * 2 + layer.weightOffset; // Convert to f16 units - - const paramsData = new Uint32Array(7); - paramsData[0] = layer.kernelSize; - paramsData[1] = layer.inChannels; - paramsData[2] = layer.outChannels; - paramsData[3] = absoluteWeightOffset; // Use absolute offset - paramsData[4] = isOutput ? 1 : 0; - paramsData[6] = (i === 0) ? 1 : 0; // is_layer_0 flag - - const paramsView = new Float32Array(paramsData.buffer); - paramsView[5] = this.blendAmount; - - const paramsBuffer = this.device.createBuffer({ - size: 28, - usage: GPUBufferUsage.UNIFORM | GPUBufferUsage.COPY_DST - }); - this.device.queue.writeBuffer(paramsBuffer, 0, paramsData); - - const cnnBG = this.device.createBindGroup({ - layout: cnnPipeline.getBindGroupLayout(0), - entries: [ - { binding: 0, resource: layerTextures[0].createView() }, - { binding: 1, resource: srcTex.createView() }, - { binding: 2, resource: dstTex.createView() }, - { binding: 3, resource: { buffer: weightsGPU } }, - { binding: 4, resource: { buffer: paramsBuffer } }, - { binding: 5, resource: this.inputTexture.createView() } - ] - }); - - const cnnPass = encoder.beginComputePass(); - cnnPass.setPipeline(cnnPipeline); - cnnPass.setBindGroup(0, cnnBG); - cnnPass.dispatchWorkgroups(Math.ceil(width / 8), Math.ceil(height / 8)); - cnnPass.end(); - - [srcTex, dstTex] = [dstTex, srcTex]; - - // Copy CNN layer output to persistent storage for visualization - // i=0: Layer 0 → layerTextures[1] - // i=1: Layer 1 → layerTextures[2], etc. - encoder.copyTextureToTexture( - { texture: srcTex }, - { texture: layerTextures[i + 1] }, - [width, height] - ); - - // Always push layer outputs for visualization (including output layer) - this.layerOutputs.push(layerTextures[i + 1]); - } - - const modeBuffer = this.device.createBuffer({ - size: 4, - usage: GPUBufferUsage.UNIFORM | GPUBufferUsage.COPY_DST - }); - this.device.queue.writeBuffer(modeBuffer, 0, new Uint32Array([this.viewMode])); - - // Store result texture and display pipeline for view mode switching - this.resultTexture = srcTex; - this.displayPipeline = displayPipeline; - this.modeBuffer = modeBuffer; - - const displayBG = this.device.createBindGroup({ - layout: displayPipeline.getBindGroupLayout(0), - entries: [ - { binding: 0, resource: srcTex.createView() }, - { binding: 1, resource: this.inputTexture.createView() }, - { binding: 2, resource: { buffer: modeBuffer } } - ] - }); - this.displayBindGroup = displayBG; - - const renderPass = encoder.beginRenderPass({ - colorAttachments: [{ - view: this.context.getCurrentTexture().createView(), - loadOp: 'clear', - storeOp: 'store' - }] - }); - renderPass.setPipeline(displayPipeline); - renderPass.setBindGroup(0, displayBG); - renderPass.draw(6); - renderPass.end(); - - this.device.queue.submit([encoder.finish()]); - - // Wait for GPU to finish before visualizing layers - await this.device.queue.onSubmittedWorkDone(); - - const t1 = performance.now(); - const mode = ['CNN Output', 'Original', 'Diff (×10)'][this.viewMode]; - this.setStatus(`GPU: ${(t1-t0).toFixed(1)}ms | ${width}×${height} | ${mode}`); - this.log(`Completed in ${(t1-t0).toFixed(1)}ms`); - - // Update layer visualization panel - this.updateLayerVizPanel(); - } - - updateLayerVizPanel() { - const panel = document.getElementById('layerViz'); - - if (!this.layerOutputs || this.layerOutputs.length === 0) { - panel.innerHTML = '<p style="color: #808080; text-align: center;">No layers to visualize</p>'; - return; - } - - // Only rebuild panel structure if layer count changed - const needsRebuild = !this.lastLayerCount || this.lastLayerCount !== this.layerOutputs.length; - - if (needsRebuild) { - let html = '<div class="layer-buttons">'; - html += `<button onclick="tester.visualizeLayer(0, 0)" id="layerBtn0_0">Static 0-3</button>`; - html += `<button onclick="tester.visualizeLayer(0, 4)" id="layerBtn0_4">Static 4-7</button>`; - - for (let i = 1; i < this.layerOutputs.length; i++) { - const label = `Layer ${i - 1}`; - html += `<button onclick="tester.visualizeLayer(${i})" id="layerBtn${i}">${label}</button>`; - } - html += `<button onclick="tester.saveCompositedLayer()" style="margin-left: 20px; background: #28a745;">Save Composited</button>`; - html += '</div>'; - - html += '<div class="layer-grid" id="layerGrid"></div>'; - html += '<div class="layer-preview"><div class="layer-view-label" id="previewLabel">Ch0</div><canvas id="previewCanvas"></canvas></div>'; - - panel.innerHTML = html; - this.log(`Layer visualization ready: ${this.layerOutputs.length} layers`); - this.recreateCanvases(); - this.lastLayerCount = this.layerOutputs.length; - } - - // Update current visualization - if (this.currentLayerIdx !== null) { - this.visualizeLayer(this.currentLayerIdx, this.currentChannelOffset || 0); - } else { - this.visualizeLayer(0, 0); - } - } - - recreateCanvases() { - const grid = document.getElementById('layerGrid'); - if (!grid) return; - - // Force removal of old canvases to clear any WebGPU contexts - const oldCanvases = grid.querySelectorAll('canvas'); - oldCanvases.forEach(canvas => { - canvas.width = 0; - canvas.height = 0; - }); - - grid.innerHTML = ''; - for (let c = 0; c < 4; c++) { - const div = document.createElement('div'); - div.className = 'layer-view'; - div.innerHTML = ` - <div class="layer-view-label" id="channelLabel${c}">Ch ${c}</div> - <canvas id="layerCanvas${c}"></canvas> - `; - div.onclick = () => this.selectChannel(c); - grid.appendChild(div); - } - this.selectedChannel = 0; - } - - async visualizeLayer(layerIdx, channelOffset = 0) { - if (!this.layerOutputs || layerIdx >= this.layerOutputs.length) { - this.log(`Cannot visualize layer ${layerIdx}: no data`, 'error'); - return; - } - - // Store current selection - this.currentLayerIdx = layerIdx; - this.currentChannelOffset = channelOffset; - - // Update button states - document.querySelectorAll('.layer-buttons button').forEach(btn => btn.classList.remove('active')); - if (layerIdx === 0) { - // Static features - const btnId = `layerBtn0_${channelOffset}`; - const btn = document.getElementById(btnId); - if (btn) btn.classList.add('active'); - } else { - const btn = document.getElementById(`layerBtn${layerIdx}`); - if (btn) btn.classList.add('active'); - } - - const layerName = layerIdx === 0 ? `Static Features (${channelOffset}-${channelOffset + 3})` : `Layer ${layerIdx - 1}`; - const layerTex = this.layerOutputs[layerIdx]; - const { width, height } = this.getDimensions(); - - // Update channel labels based on layer type - // Static features (layerIdx=0): 8 channels split into two views - // CNN layers (layerIdx≥1): 4 channels per layer - const staticLabels = [ - ['Ch0 (p0)', 'Ch1 (p1)', 'Ch2 (p2)', 'Ch3 (p3)'], - ['Ch4 (uv_x)', 'Ch5 (uv_y)', 'Ch6 (sin10_x)', 'Ch7 (bias)'] - ]; - const channelLabels = layerIdx === 0 - ? staticLabels[channelOffset / 4] - : ['Ch0', 'Ch1', 'Ch2', 'Ch3']; - - for (let c = 0; c < 4; c++) { - const label = document.getElementById(`channelLabel${c}`); - if (label) label.textContent = channelLabels[c]; - } - - // Create layer viz pipeline if needed - if (!this.layerVizPipeline) { - this.layerVizPipeline = this.device.createRenderPipeline({ - layout: 'auto', - vertex: { - module: this.device.createShaderModule({ code: LAYER_VIZ_SHADER }), - entryPoint: 'vs_main' - }, - fragment: { - module: this.device.createShaderModule({ code: LAYER_VIZ_SHADER }), - entryPoint: 'fs_main', - targets: [{ format: this.format }] - } - }); - this.log('Created layer visualization pipeline'); - } - - // Render each channel to its canvas - for (let c = 0; c < 4; c++) { - const canvas = document.getElementById(`layerCanvas${c}`); - if (!canvas) { - this.log(`Canvas layerCanvas${c} not found`, 'error'); - continue; - } - - // Set canvas size BEFORE getting context - canvas.width = width; - canvas.height = height; - - const ctx = canvas.getContext('webgpu'); - if (!ctx) { - this.log(`Failed to get WebGPU context for channel ${c}`, 'error'); - continue; - } - - try { - ctx.configure({ device: this.device, format: this.format }); - } catch (e) { - this.log(`Failed to configure canvas ${c}: ${e.message}`, 'error'); - continue; - } - - const vizScale = 1.0; // Always 1.0, shader clamps to [0,1] - const paramsBuffer = this.device.createBuffer({ - size: 8, - usage: GPUBufferUsage.UNIFORM | GPUBufferUsage.COPY_DST - }); - // Use channel index with offset for static features - const actualChannel = channelOffset + c; - const paramsData = new Float32Array([actualChannel, vizScale]); - this.device.queue.writeBuffer(paramsBuffer, 0, paramsData); - - const bindGroup = this.device.createBindGroup({ - layout: this.layerVizPipeline.getBindGroupLayout(0), - entries: [ - { binding: 0, resource: layerTex.createView() }, - { binding: 1, resource: { buffer: paramsBuffer } } - ] - }); - - const encoder = this.device.createCommandEncoder(); - const renderPass = encoder.beginRenderPass({ - colorAttachments: [{ - view: ctx.getCurrentTexture().createView(), - loadOp: 'clear', - clearValue: { r: 1.0, g: 0.0, b: 1.0, a: 1.0 }, // Magenta clear for debugging - storeOp: 'store' - }] - }); - - renderPass.setPipeline(this.layerVizPipeline); - renderPass.setBindGroup(0, bindGroup); - renderPass.draw(6); - renderPass.end(); - - this.device.queue.submit([encoder.finish()]); - } - - // Wait for all renders to complete - await this.device.queue.onSubmittedWorkDone(); - - // Update active channel highlighting and preview - this.updateChannelSelection(); - await this.renderChannelPreview(); - } - - selectChannel(channelIdx) { - this.selectedChannel = channelIdx; - this.updateChannelSelection(); - this.renderChannelPreview(); - } - - updateChannelSelection() { - const grid = document.getElementById('layerGrid'); - if (!grid) return; - - const views = grid.querySelectorAll('.layer-view'); - views.forEach((view, idx) => { - view.classList.toggle('active', idx === this.selectedChannel); - }); - } - - async renderChannelPreview() { - const previewCanvas = document.getElementById('previewCanvas'); - const previewLabel = document.getElementById('previewLabel'); - if (!previewCanvas || !this.device) return; - - const { width, height } = this.getDimensions(); - previewCanvas.width = width; - previewCanvas.height = height; - - const ctx = previewCanvas.getContext('webgpu'); - if (!ctx) return; - - try { - ctx.configure({ device: this.device, format: this.format }); - } catch (e) { - return; - } - - // Update label - const channelLabel = document.getElementById(`channelLabel${this.selectedChannel}`); - if (channelLabel && previewLabel) { - previewLabel.textContent = channelLabel.textContent; - } - - // Render selected channel - const layerIdx = this.currentLayerIdx; - const channelOffset = this.currentChannelOffset; - const layerTex = this.layerOutputs[layerIdx]; - if (!layerTex) return; - - // Always 1.0, shader clamps to [0,1] - show exact layer values - const vizScale = 1.0; - const actualChannel = channelOffset + this.selectedChannel; - - const paramsBuffer = this.device.createBuffer({ - size: 8, - usage: GPUBufferUsage.UNIFORM | GPUBufferUsage.COPY_DST - }); - const paramsData = new Float32Array([actualChannel, vizScale]); - this.device.queue.writeBuffer(paramsBuffer, 0, paramsData); - - const bindGroup = this.device.createBindGroup({ - layout: this.layerVizPipeline.getBindGroupLayout(0), - entries: [ - { binding: 0, resource: layerTex.createView() }, - { binding: 1, resource: { buffer: paramsBuffer } } - ] - }); - - const encoder = this.device.createCommandEncoder(); - const renderPass = encoder.beginRenderPass({ - colorAttachments: [{ - view: ctx.getCurrentTexture().createView(), - loadOp: 'clear', - storeOp: 'store' - }] - }); - - renderPass.setPipeline(this.layerVizPipeline); - renderPass.setBindGroup(0, bindGroup); - renderPass.draw(6); - renderPass.end(); - - this.device.queue.submit([encoder.finish()]); - } - - visualizeWeights(cnnLayerIdx) { - const layer = this.weights.layers[cnnLayerIdx]; - if (!layer) { - this.log(`Layer ${cnnLayerIdx} not found`, 'error'); - return; - } - - // Update button states - document.querySelectorAll('#weightsLayerButtons button').forEach(btn => btn.classList.remove('active')); - const btn = document.getElementById(`weightsBtn${cnnLayerIdx}`); - if (btn) btn.classList.add('active'); - - const { kernelSize, inChannels, outChannels, weightOffset, min, max } = layer; - - const canvas = document.getElementById('weightsCanvas'); - const ctx = canvas.getContext('2d', { willReadFrequently: false }); - - // 1 pixel per weight, show all input channels horizontally - const width = inChannels * kernelSize; - const height = outChannels * kernelSize; - - canvas.width = width; - canvas.height = height; - - ctx.fillStyle = '#1a1a1a'; - ctx.fillRect(0, 0, width, height); - - // Stack output channels vertically - for (let outCh = 0; outCh < outChannels; outCh++) { - const yOffset = outCh * kernelSize; - - for (let inCh = 0; inCh < inChannels; inCh++) { - const xOffset = inCh * kernelSize; - - for (let ky = 0; ky < kernelSize; ky++) { - for (let kx = 0; kx < kernelSize; kx++) { - const spatialIdx = ky * kernelSize + kx; - const wIdx = weightOffset + - outCh * inChannels * kernelSize * kernelSize + - inCh * kernelSize * kernelSize + - spatialIdx; - - const weight = this.getWeightValue(wIdx); - const normalized = (weight - min) / (max - min); - const intensity = Math.floor(normalized * 255); - - ctx.fillStyle = `rgb(${intensity}, ${intensity}, ${intensity})`; - ctx.fillRect(xOffset + kx, yOffset + ky, 1, 1); - } - } - } - } - } - - getWeightValue(idx) { - const pairIdx = Math.floor(idx / 2); - const packed = this.weights.weights[pairIdx]; - const unpacked = this.unpackF16(packed); - return (idx % 2 === 0) ? unpacked[0] : unpacked[1]; - } - - toggleWeightsInfo() { - const panel = document.getElementById('weightsInfoPanel'); - const toggle = document.getElementById('weightsInfoToggle'); - panel.classList.toggle('collapsed'); - toggle.textContent = panel.classList.contains('collapsed') ? '▶' : '▼'; - } - - updateDisplay() { - if (!this.displayPipeline || !this.displayBindGroup) return; - - this.device.queue.writeBuffer(this.modeBuffer, 0, new Uint32Array([this.viewMode])); - - const encoder = this.device.createCommandEncoder(); - const renderPass = encoder.beginRenderPass({ - colorAttachments: [{ - view: this.context.getCurrentTexture().createView(), - loadOp: 'clear', - storeOp: 'store' - }] - }); - renderPass.setPipeline(this.displayPipeline); - renderPass.setBindGroup(0, this.displayBindGroup); - renderPass.draw(6); - renderPass.end(); - - this.device.queue.submit([encoder.finish()]); - } - - async savePNG() { - if (!this.image && !this.isVideo) { - this.log('No image loaded', 'error'); - return; - } - - if (!this.resultTexture) { - this.log('No result to save', 'error'); - return; - } - - try { - const { width, height } = this.getDimensions(); - - // GPU readback from result texture - const bytesPerRow = width * 16; // 4×u32 per pixel - const paddedBytesPerRow = Math.ceil(bytesPerRow / 256) * 256; - const bufferSize = paddedBytesPerRow * height; - - const stagingBuffer = this.device.createBuffer({ - size: bufferSize, - usage: GPUBufferUsage.COPY_DST | GPUBufferUsage.MAP_READ - }); - - const encoder = this.device.createCommandEncoder(); - encoder.copyTextureToBuffer( - { texture: this.resultTexture }, - { buffer: stagingBuffer, bytesPerRow: paddedBytesPerRow, rowsPerImage: height }, - { width, height, depthOrArrayLayers: 1 } - ); - this.device.queue.submit([encoder.finish()]); - - await stagingBuffer.mapAsync(GPUMapMode.READ); - const mapped = new Uint8Array(stagingBuffer.getMappedRange()); - - // Unpack f16 to RGBA8 - const pixels = new Uint8Array(width * height * 4); - for (let y = 0; y < height; y++) { - const rowOffset = y * paddedBytesPerRow; - for (let x = 0; x < width; x++) { - const pixelOffset = rowOffset + x * 16; - const data = new Uint32Array(mapped.buffer, mapped.byteOffset + pixelOffset, 4); - - // Unpack f16 (first 4 channels only) - const unpack = (u32, idx) => { - const h = (idx === 0) ? (u32 & 0xFFFF) : ((u32 >> 16) & 0xFFFF); - const sign = (h >> 15) & 1; - const exp = (h >> 10) & 0x1F; - const frac = h & 0x3FF; - if (exp === 0) return 0; - if (exp === 31) return sign ? 0 : 255; - const e = exp - 15; - const val = (1 + frac / 1024) * Math.pow(2, e); - return Math.max(0, Math.min(255, Math.round(val * 255))); - }; - - const outIdx = (y * width + x) * 4; - pixels[outIdx + 0] = unpack(data[0], 0); // R - pixels[outIdx + 1] = unpack(data[0], 1); // G - pixels[outIdx + 2] = unpack(data[1], 0); // B - pixels[outIdx + 3] = 255; // A - } - } - - stagingBuffer.unmap(); - stagingBuffer.destroy(); - - // Create blob from pixels - const canvas = document.createElement('canvas'); - canvas.width = width; - canvas.height = height; - const ctx = canvas.getContext('2d'); - const imageData = new ImageData(new Uint8ClampedArray(pixels), width, height); - ctx.putImageData(imageData, 0, 0); - - const blob = await new Promise(resolve => canvas.toBlob(resolve, 'image/png')); - const url = URL.createObjectURL(blob); - const a = document.createElement('a'); - const mode = ['cnn', 'original', 'diff'][this.viewMode]; - a.href = url; - a.download = `output_${width}x${height}_${mode}.png`; - a.click(); - URL.revokeObjectURL(url); - - this.log(`Saved PNG: ${a.download}`); - this.setStatus(`Saved: ${a.download}`); - } catch (err) { - this.log(`Failed to save PNG: ${err.message}`, 'error'); - this.setStatus(`Save failed: ${err.message}`, true); - } - } - - async saveCompositedLayer() { - if (!this.currentLayerIdx) { - this.log('No layer selected for compositing', 'error'); - return; - } - - try { - const canvases = []; - for (let i = 0; i < 4; i++) { - const canvas = document.getElementById(`layerCanvas${i}`); - if (!canvas) { - this.log(`Canvas layerCanvas${i} not found`, 'error'); - return; - } - canvases.push(canvas); - } - - const width = canvases[0].width; - const height = canvases[0].height; - const compositedWidth = width * 4; - - // Create composited canvas - const compositedCanvas = document.createElement('canvas'); - compositedCanvas.width = compositedWidth; - compositedCanvas.height = height; - const ctx = compositedCanvas.getContext('2d'); - - // Composite horizontally - for (let i = 0; i < 4; i++) { - ctx.drawImage(canvases[i], i * width, 0); - } - - // Convert to grayscale - const imageData = ctx.getImageData(0, 0, compositedWidth, height); - const pixels = imageData.data; - for (let i = 0; i < pixels.length; i += 4) { - const gray = 0.299 * pixels[i] + 0.587 * pixels[i + 1] + 0.114 * pixels[i + 2]; - pixels[i] = pixels[i + 1] = pixels[i + 2] = gray; - } - ctx.putImageData(imageData, 0, 0); - - // Save as PNG - const blob = await new Promise(resolve => compositedCanvas.toBlob(resolve, 'image/png')); - const url = URL.createObjectURL(blob); - const a = document.createElement('a'); - a.href = url; - a.download = `composited_layer${this.currentLayerIdx - 1}_${compositedWidth}x${height}.png`; - a.click(); - URL.revokeObjectURL(url); - - this.log(`Saved composited layer: ${a.download}`); - this.setStatus(`Saved: ${a.download}`); - } catch (err) { - this.log(`Failed to save composited layer: ${err.message}`, 'error'); - this.setStatus(`Compositing failed: ${err.message}`, true); - } - } -} - -const tester = new CNNTester(); - -// Load default weights on startup -(async () => { - try { - const binaryString = atob(DEFAULT_WEIGHTS_B64); - const bytes = new Uint8Array(binaryString.length); - for (let i = 0; i < binaryString.length; i++) { - bytes[i] = binaryString.charCodeAt(i); - } - await tester.loadWeights({ name: 'default.bin', arrayBuffer: () => Promise.resolve(bytes.buffer) }); - tester.log('Loaded default weights'); - } catch (err) { - tester.log(`Failed to load default weights: ${err.message}`, 'error'); - } -})(); - -function setupDropZone(id, callback) { - const zone = document.getElementById(id); - ['dragenter', 'dragover', 'dragleave', 'drop'].forEach(e => { - zone.addEventListener(e, ev => { ev.preventDefault(); ev.stopPropagation(); }); - }); - ['dragenter', 'dragover'].forEach(e => zone.addEventListener(e, () => zone.classList.add('active'))); - ['dragleave', 'drop'].forEach(e => zone.addEventListener(e, () => zone.classList.remove('active'))); - zone.addEventListener('drop', e => { - const file = e.dataTransfer.files[0]; - if (file) callback(file).catch(err => { - zone.classList.add('error'); - tester.setStatus(err.message, true); - tester.log(err.message, 'error'); - setTimeout(() => zone.classList.remove('error'), 2000); - }); - }); -} - -// Whole window drop for PNG images and videos -const mainArea = document.getElementById('mainDrop'); -['dragenter', 'dragover', 'dragleave', 'drop'].forEach(e => { - mainArea.addEventListener(e, ev => { ev.preventDefault(); ev.stopPropagation(); }); -}); -['dragenter', 'dragover'].forEach(e => mainArea.addEventListener(e, () => mainArea.classList.add('drop-active'))); -['dragleave', 'drop'].forEach(e => mainArea.addEventListener(e, () => mainArea.classList.remove('drop-active'))); -mainArea.addEventListener('drop', e => { - const file = e.dataTransfer.files[0]; - if (file) { - if (file.type.startsWith('image/')) { - tester.loadImage(file).catch(err => { - tester.setStatus(err.message, true); - tester.log(err.message, 'error'); - }); - } else if (file.type.startsWith('video/')) { - tester.loadVideo(file).catch(err => { - tester.setStatus(err.message, true); - tester.log(err.message, 'error'); - }); - } - } -}); - -// Weights drop zone -setupDropZone('weightsDrop', f => tester.loadWeights(f)); - -// Weights file input -document.getElementById('weightsFile').addEventListener('change', e => { - const file = e.target.files[0]; - if (file) { - tester.loadWeights(file).catch(err => { - tester.setStatus(err.message, true); - tester.log(err.message, 'error'); - }); - } -}); - -document.getElementById('blend').addEventListener('input', e => { - tester.blendAmount = parseFloat(e.target.value); - document.getElementById('blendValue').textContent = e.target.value; - if ((tester.image || tester.isVideo) && tester.weights) { - tester.log(`Blend changed to ${e.target.value}`); - tester.run(); - } -}); - -document.getElementById('depth').addEventListener('input', e => { - tester.depth = parseFloat(e.target.value); - document.getElementById('depthValue').textContent = e.target.value; - if ((tester.image || tester.isVideo) && tester.weights) tester.run(); -}); - -document.getElementById('mipLevel').addEventListener('change', e => { - tester.mipLevel = parseInt(e.target.value); - tester.log(`Mip level changed to ${e.target.value}`); - if ((tester.image || tester.isVideo) && tester.weights) tester.run(); -}); - -document.getElementById('playPauseBtn').addEventListener('click', () => tester.togglePlayPause()); -document.getElementById('stepBackBtn').addEventListener('click', () => tester.stepFrame(-1)); -document.getElementById('stepForwardBtn').addEventListener('click', () => tester.stepFrame(1)); -document.getElementById('savePngBtn').addEventListener('click', () => tester.savePNG()); - -document.addEventListener('keydown', e => { - if (e.code === 'Space') { - e.preventDefault(); - if (tester.viewMode === 1) { - tester.viewMode = 0; - } else { - tester.viewMode = 1; - } - const modeName = ['CNN Output', 'Original', 'Diff (×10)'][tester.viewMode]; - if ((tester.image || tester.isVideo) && tester.weights) { - tester.log(`View mode: ${modeName}`); - tester.updateDisplay(); - const width = tester.isVideo ? tester.video.videoWidth : tester.image.width; - const height = tester.isVideo ? tester.video.videoHeight : tester.image.height; - tester.setStatus(`${width}×${height} | ${modeName}`); - } - } else if (e.code === 'KeyD') { - e.preventDefault(); - if (tester.viewMode === 2) { - tester.viewMode = 0; - } else { - tester.viewMode = 2; - } - const modeName = ['CNN Output', 'Original', 'Diff (×10)'][tester.viewMode]; - if ((tester.image || tester.isVideo) && tester.weights) { - tester.log(`View mode: ${modeName}`); - tester.updateDisplay(); - const width = tester.isVideo ? tester.video.videoWidth : tester.image.width; - const height = tester.isVideo ? tester.video.videoHeight : tester.image.height; - tester.setStatus(`${width}×${height} | ${modeName}`); - } - } -}); - </script> -</body> -</html> |
