Same prompt, two backends. Mix Ollama API models and local WebGPU models side by side — compare quality and speed across different backends and sizes.
OLLAMA_ORIGINS='*' ollama serve