Thursday, 07 May 2026

Anthropic secures SpaceX compute to double Claude limits, Coinbase cuts 14% of staff in AI-native restructure, and the line between vibe coding and professional engineering blurs

Today's Lead

Anthropic

Anthropic Secures SpaceX Compute Partnership, Immediately Doubles Claude Code Rate Limits

Anthropic announced a compute partnership with SpaceX/xAI granting access to over 300 megawatts of capacity — some 220,000+ NVIDIA GPUs — from Colossus 1, ramping within days. The deal's immediate effect is concrete: Claude Code's five-hour rate limits are doubled for Pro, Max, Team, and Enterprise plans, peak-hour throttling is removed for Pro and Max, and Opus API rate limits are substantially increased. The backstory matters as much as the announcement: usage reportedly grew 80× faster than anticipated, which is what caused the rate limits in the first place — this was a capacity crisis, not a pricing decision, and the SpaceX deal is the first major attempt to address it. The deeper implication is structural: inference, not just training, has become a frontier bottleneck, and compute markets are now fluid enough that one frontier lab can rent datacenter capacity from another's affiliated infrastructure. Anthropic renting from SpaceX/xAI — an arrangement that required Elon Musk's sign-off and arrived just as his lawsuit against OpenAI went to trial — adds an unusually political dimension to what is ostensibly a procurement announcement. The battle between frontier AI labs is increasingly being fought at the level of compute, rate limits, and agent ergonomics, not just benchmark leaderboards.

Read →

Also today

LeadDev

Coinbase Cuts 14% of Staff, Flattens Management in AI-Native Restructure

Coinbase announced a 14% global workforce reduction alongside a fundamental reorganization: management is capped at five layers below CEO/COO, and traditional manager roles are being replaced by 'player-coaches' who combine oversight with hands-on technical work. The more striking concept is the 'AI-native pod' — a structure in which a single engineer may direct AI agents handling engineering, design, and product work simultaneously. CEO Brian Armstrong framed this as rebuilding Coinbase 'as an intelligence, with humans around the edge aligning it.' This is one of the most explicit acknowledgments from a major public company that AI is the proximate cause of a workforce reduction, rather than an implicit backdrop. The organizational model Armstrong is describing — where individual contributors expand their scope by delegating to AI agents rather than headcount — represents a serious hypothesis about how companies will be structured in five years. Whether it works depends on whether AI agents can reliably hold context across the cross-functional coordination that traditional teams handle, a question that remains empirically open.

Read →

Simon Willison

Vibe Coding and Agentic Engineering Are Starting to Converge — Uncomfortably

Simon Willison, who coined the distinction between 'vibe coding' (non-programmers blindly accepting AI output) and 'agentic engineering' (professionals using AI as an amplifier), now admits the boundary is blurring in his own practice. The discomfort: as AI coding agents become reliably correct on routine tasks, even experienced engineers increasingly skip reviewing their output — not from carelessness but because the track record justifies it. Willison's analogy is apt: he treats Claude Code the way an engineering manager treats a trusted team — using the service without auditing every line, until something breaks. The risk he names is 'normalization of deviance': each successful unreviewed deployment incrementally erodes the instinct to check, setting up a future failure that the engineer has quietly trained themselves not to catch. His proposed reframe is useful: the value of a vibe-coded project is now less about test coverage and documentation and more about whether anyone has actually used it daily — proof of real-world exercising that synthetic thoroughness can't replicate.

Read →

GitHub Blog

GitHub Applies Compiler Theory to Validate Non-Deterministic Agent Behavior

GitHub's engineering team published a detailed technical post on validating GitHub Copilot Coding Agent's Computer Use behavior — specifically the problem where an agent completes a task successfully but CI still flags failure because a loading screen appeared that wasn't in the test script. Their solution borrows from compiler theory: build a directed graph from 2–10 successful execution traces, apply dominator analysis to identify which states are mathematically required on every successful path versus incidental noise, then validate new runs by checking only the essential states appear in order. A three-tier equivalence check (perceptual hashing, structural similarity, LLM semantic analysis) handles the hard problem of deciding whether two screenshots represent the same logical state. The results are striking: 100% accuracy versus 82.2% for agent self-assessment, and — critically — 0% F1-score for agents judging their own 'not a bug' scenarios, meaning agents cannot reliably grade their own homework in non-deterministic environments. This paper represents serious applied engineering on a problem every team deploying AI agents in CI/CD pipelines is about to hit.

Read →

SQLite

The Library of Congress Recommends SQLite as an Archival Storage Format

The Library of Congress has designated SQLite as a Recommended Storage Format for digital preservation — one of only four formats in the list alongside XML, JSON, and CSV. SQLite meets all seven of the LoC's key preservation criteria: open and documented format, wide adoption across ecosystems, ability to examine data directly, internal metadata storage, no proprietary dependencies, freedom from patents, and no encryption or access restrictions. The practical implication is significant for institutions: agencies and archives that accept the LoC recommendation now have a defensible procurement justification for SQLite as a long-term data store. The deeper point is about what SQLite's architecture was already optimized for: a single-file format that carries its schema, data, and metadata together, designed to remain readable without the application that created it. The endorsement is a quiet validation of the design philosophy that has made SQLite the most widely deployed database engine on earth.

Read →

val.town Blog

val.town Migrated Authentication Twice — From Supabase to Clerk to Better Auth

val.town's engineering blog documents two successive authentication migrations and the specific failures that drove each. Moving from Supabase Auth to Clerk in 2023 solved early scaling concerns, but Clerk's 5-requests-per-second rate limit became a ceiling for a platform with social features — displaying user avatars required a parallel users table and custom webhooks to work around the constraint. More critically: Clerk operated as a hard dependency, meaning any Clerk outage made val.town completely unusable even for already-authenticated users. The move to Better Auth eliminated the vendor single point of failure, enabling self-hosted session management and removing the external uptime dependency from the critical path. The lesson is transferable: managed authentication services are optimized for simple applications and their constraints only become visible at scale or with unconventional usage patterns. The specific failure mode — a rate limit causing architectural workarounds that introduce their own reliability risks — is exactly the kind of debt that accrues invisibly until it doesn't.

Read →

alexsci.com

An HTTP Header Was Making Time.gov Systematically Wrong

NIST's time.gov synchronizes browsers with atomic clocks by measuring a single HTTP round-trip: the page sends a request, records the server timestamp in the response, and computes offset from the measured latency. The bug: a 'Connection: close' header forced a fresh TCP/TLS handshake for every request, tripling the number of round-trips. Since the server timestamp is captured after only part of the handshake while the client measured the full three-round-trip cost, the offset calculation was systematically wrong — by approximately one full round-trip. The irony is rich: the authoritative U.S. time reference was serving subtly inaccurate time because of an incidental infrastructure header, not a clock error. The broader lesson is about invisible assumptions in distributed systems: the JavaScript measuring latency assumed the server would reuse an existing connection, an assumption that held until some infrastructure change (load balancer, deployment update, firewall rule) quietly inserted the header. Production reliability depends on understanding every layer of the stack, especially the layers your application code doesn't control.

Read →

stvn.sh

Programming Still Sucks — and AI Isn't the Problem

A widely-read post (290 points on HN) argues that software engineering's current dysfunction is driven by corporate cost-cutting, not technological limits — and that AI is being used as cover for decisions that would have happened anyway. The three specific failures: the elimination of entry-level positions that previously built institutional knowledge, the collapse of code review culture in favor of shipping metrics, and the resulting loss of the shared understanding that made large codebases maintainable. The author's sharpest point is about the AI job displacement narrative: companies claiming AI makes junior roles unnecessary are actually using AI as a justification for a staffing decision they wanted to make for cost reasons, with the predictable consequence that the pipeline that trained senior engineers is now broken. This matters independently of whether AI will eventually automate software development entirely — the skills attrition and institutional knowledge loss are happening now, and the effects will compound regardless of how capable the models become.

Read →

Proton

Proton Launches End-to-End Encrypted Video Conferencing

Proton has launched Proton Meet, a video conferencing service using the Messaging Layer Security protocol to provide end-to-end encryption for calls, chats, and screen shares — making it inaccessible even to Proton itself. The key differentiator from Zoom, Google Meet, and Teams is the explicit prohibition on data harvesting combined with Swiss legal jurisdiction, which provides meaningful structural protection against U.S. CLOUD Act surveillance requests rather than a policy promise. The free tier supports up to 50 attendees with calendar integration and real-time encrypted messaging. Proton's play is to be the privacy-complete alternative to Google Workspace — it already offers email (ProtonMail), calendar, cloud storage (Proton Drive), and VPN (ProtonVPN), and now video. For organizations under GDPR or operating in sensitive contexts, the value proposition is not just features but a coherent data residency and access story across the full collaboration stack.

Read →

jdx.dev

Mise Maintainer Leaves Figma to Go Full-Time on Open Source

Jeff Dickey (jdx), the maintainer behind mise — a polyglot dev tools manager with 27,000+ GitHub stars and the 10th most downloaded Homebrew formula — left his role at Figma to work on open source full-time through his company en.dev. The economics are stark: current sponsorship revenue of roughly $600/month covers a small fraction of what's needed, requiring a combination of corporate sponsorships, memberships, and consulting to make it viable. The project portfolio is substantial — mise, aube (a new Node.js package manager), hk, pitchfork, and others — and PR review backlogs were becoming untenable alongside full-time employment. The story is a clean example of the open source sustainability problem: tools that become load-bearing infrastructure for tens of thousands of developers often generate more GitHub stars than dollars, and the maintainer has to construct a business model from scratch or burn out. The indie open source path has become more viable than it was five years ago, but the gap between 'widely used' and 'financially sustainable' remains substantial.

Read →