The Daily Claw Issue #0034 - AutoKernel keeps GPU kernels awake
AutoKernel keeps GPU kernels awake
AutoKernel is the kind of agentic system that makes your GPU stack think twice before you commit another manual patch. The service keeps an eye on every experiment, profiles each kernel, and summons a Claude/Codex pair to rewrite bottlenecks until the end-to-end benchmark improves. Its pipeline—profile → extract → benchmark → verify—runs on a ~90-second cadence, meaning the overnight batch of 320+ experiments never lets a dud ship. If you care about winning the latency arms race, treat the AutoKernel launch thread as a playbook for running models in production: keep GPU kernel templates (matmul, softmax, flash attention, fused MLP, rotary, etc.) under continual autotuning, and delete the edits that don’t pass roofline analysis.
Gemini Embedding 2 unifies every modality
The same mindset applies to the primitives you build on top of those kernels. Google’s Gemini Embedding 2 announcement folds text, image, video, and PDF into a single endpoint that understands up to 8,192 tokens, six PNG/JPEG inputs, 120 seconds of video, or six pages of documents per call. You can tune the output dimension (3,072 → 1,536 → 768) without shattering the vector space that powers your RAG stack, so upgrade your semantic search or multimodal index knowing the new endpoint speaks every modality you already store.
Source maps Revision 3 keeps every debugger honest
If bundler noise is keeping your debug sessions dirty, Bloomberg’s deep dive on Source Maps Revision 3 is a reminder that standards move faster than whatever custom mapping logic your team is shipping. The new JSON schema (version, sources, names, mappings, optional ignoreList) lets you handle 200k–2M character bundles without per-column bloat, and proposals for ignoreLists plus inline sources mean every toolchain—from bundler to debugger—gets the same expectation of accuracy. Audit every build step today so you don’t end up chasing a distortion between Chrome, Safari, and your observability dashboards.
Quick hits

- Amazon now insists on senior sign-off for AI-assisted changes; build approval chains and rollout playbooks before one of your copilots makes a config change that takes the site offline, especially after the latest outages.
- Hume open-sourced TADA, a TTS engine that runs at 0.09x real time, hits zero hallucinations across LibriTTSR, and lets you bundle a deterministic voice right alongside your on-device agents via their blog post.
- InsForge now ships auth, storage, edge functions, tool calling, and a model gateway so agents can build and deploy a fullstack product from prompt to launch; read their launch details at insforge.dev.
- DevToolkit API keeps a grab bag of JWT decoders, UUID generators, SSL checks, redirect tracers, and webhook inspectors ready without a login; bookmark the utility console for your next “does this cert chain work?” moment.
- Detect Anything provides a Colab pipeline plus API for auto-labeling YOLO datasets, exports training weights, and comes with OpenAPI docs, making it easier to bootstrap domain-specific detectors before hiring labelers; dip into the GitHub repo.