AI News

269

Claude Code users hitting usage limits 'way faster than expected'

Claude Code users hitting usage limits 'way faster than expected'
Mastodon +7 sources mastodon
anthropicclaude
Anthropic’s Claude Code platform is throttling developers faster than anticipated, prompting a wave of complaints across Reddit, GitHub and tech forums. Users report that both free and paid tiers exhaust their token quotas within hours of a typical session, a stark contrast to the multi‑day usage windows advertised in the service’s launch notes. One Reddit commenter highlighted that a $100‑per‑month subscription, which should have afforded a substantially higher allowance, ran out “much later” than a free account, suggesting the throttling is indiscriminate. The surge in limit breaches follows a series of performance setbacks reported earlier this month, including the “Claude Code Down” outage and the February update that rendered the tool “unusable for complex engineering tasks.” As we reported on April 6, users were already experimenting with work‑arounds to stretch their quotas, but the current drain appears to be a systemic issue rather than isolated misconfigurations. Anthropic has publicly acknowledged the problem, stating that the team is “actively investigating” and that a fix is a top priority. The company’s response is critical because Claude Code is positioned as a flagship product for AI‑assisted software development, and rapid quota depletion threatens its credibility among enterprise customers who rely on predictable compute budgeting. Moreover, the episode underscores a broader industry challenge: balancing generous usage caps with the high compute costs of large language models, especially when they are embedded in IDE‑style environments that encourage continuous prompting. What to watch next: Anthropic is expected to release a detailed post‑mortem and revised quota policy within the next week. Developers should monitor the official status page for any temporary relief measures, such as increased token limits or tier‑specific exemptions. The incident also raises the question of whether Anthropic will introduce a metered‑pay‑as‑you‑go model to replace the current flat‑rate subscriptions, a shift that could reshape pricing across the AI‑coding market.
170

💻 I took several completely independent datasets and "pitted" them against each other. One of the re

💻 I took several completely independent datasets and "pitted" them against each other. One of the re
Mastodon +6 sources mastodon
huggingface
A data‑driven experiment posted this week shows a stark, quantifiable link between the built environment and local heat levels. The author combined three publicly available datasets – high‑resolution satellite imagery, a pretrained computer‑vision model that tags “concrete” features such as roads, buildings and parking lots, and thermal‑sensor readings from a network of ground‑based stations – and ran them side by side for dozens of neighbourhoods across Scandinavia and Central Europe. The resulting chart, highlighted in the post, reveals a near‑linear rise in surface temperature as the proportion of concrete‑identified pixels increases. In the hottest sampled districts, concrete cover exceeds 70 % and recorded temperatures are up to 5 °C above the regional average. The finding matters because it provides a low‑cost, AI‑enabled method for mapping urban heat islands in real time. Traditional heat‑island studies rely on sparse weather stations or expensive aerial surveys; the new approach leverages existing open‑source imagery and a generic object‑detection model, making it scalable to any city with satellite coverage. Policymakers can therefore pinpoint hotspots, prioritize greening projects, and evaluate the cooling impact of new construction before ground is broken. The work also underscores a broader trend: machine‑learning models trained on unrelated tasks (here, object detection) can be repurposed as environmental sensors when paired with complementary data streams. What to watch next is the translation of this proof‑of‑concept into municipal planning tools. Several Nordic municipalities have already expressed interest in pilot programmes that integrate the model’s outputs with GIS platforms for zoning decisions. Meanwhile, researchers are testing whether the same methodology can flag other climate‑relevant features, such as tree canopy loss or reflective roof adoption. If the early results hold, AI‑driven “data‑pitting” could become a staple of climate‑smart urban design.
123

Iran threatens OpenAI’s Stargate data center in Abu Dhabi

Iran threatens OpenAI’s Stargate data center in Abu Dhabi
Mastodon +8 sources mastodon
openai
Iran’s Islamic Revolutionary Guard Corps (IRGC) has escalated its campaign against OpenAI by publishing a new video that threatens to “totally destroy” the company’s planned $30 billion Stargate data centre in Abu Dhabi. The footage, released on state‑run channels, pairs satellite imagery of the 1 GW facility with a warning that any U.S. strike on Iranian power infrastructure will trigger a retaliatory attack on the AI hub. The message is framed as a direct response to what Tehran calls the “Tangerine Tyrant,” a reference to recent U.S. cyber‑operations targeting Iranian energy assets. The Stargate project, a joint venture involving OpenAI, Microsoft and regional cloud providers, is intended to become a cornerstone of global AI compute, housing thousands of GPUs that will power next‑generation models for both commercial and research use. Its location in the United Arab Emirates gives the venture strategic distance from the United States while still providing low‑latency connectivity to Asian and European markets. A successful IRGC strike would not only cripple OpenAI’s compute capacity but also signal that critical AI infrastructure is now a frontline in geopolitical rivalries. As we reported on 6 April, Iran had already threatened the centre, but the new conditional threat marks a shift from blanket intimidation to a tit‑for‑tat stance linked to U.S. actions. The development raises immediate questions about the security protocols surrounding the site, the feasibility of hardening a 1 GW data centre against missile or drone attacks, and whether OpenAI will diversify its compute assets further away from contested regions. What to watch next: statements from the U.S. Department of Defense and the State Department on any planned strikes; OpenAI’s response, including possible relocation of hardware or acceleration of redundancy plans; diplomatic engagement between the UAE and Tehran; and the broader impact on the emerging market for sovereign AI data centres, which could see heightened insurance costs and a re‑assessment of risk‑adjusted investment.
96

Google AI Just Made ChatGPT and Claude Obsolete (+ 13 Top AI Updates) https://www. yayafa.com/27

Google AI Just Made ChatGPT and Claude Obsolete  (+ 13 Top AI Updates)   https://www.  yayafa.com/27
Mastodon +8 sources mastodon
agentsclaudegoogle
Google unveiled its latest Gemini model, dubbed “Gemini Ultra,” and positioned it as a generative‑AI system that outperforms both OpenAI’s ChatGPT‑4 and Anthropic’s Claude 3 across a suite of benchmark tests. The announcement, made at the company’s AI Summit in Tokyo, highlighted a 15‑point lead on the MMLU reasoning exam, a 20‑percent reduction in hallucinations on factual queries, and multimodal capabilities that let developers feed text, images and code into a single prompt. Google’s engineers also demonstrated real‑time tool use, where Gemini Ultra autonomously calls APIs, drafts spreadsheets and even writes short‑form video scripts, a step the company calls “agentic AI.” The claim matters because it reshapes the competitive landscape that has been dominated by ChatGPT’s rapid adoption and Claude’s niche appeal among developers. Google’s integration of Gemini Ultra into Search, Workspace and the Cloud AI platform means enterprises can tap the model without leaving their existing ecosystems, potentially accelerating migration away from OpenAI’s API and Anthropic’s limited‑access offerings. The move also arrives as Claude users have been hitting usage caps and experiencing downtime, issues we covered on April 6 and 7, underscoring demand for a more reliable, high‑throughput alternative. What to watch next is the rollout schedule and pricing model. Google said the API will be beta‑available to select partners next month, with a broader launch slated for Q4. Analysts will be tracking performance on domain‑specific tasks such as medical coding and legal brief drafting, where OpenAI and Anthropic have recently claimed headway. Equally important will be regulatory scrutiny in Europe and the Nordics, where data‑privacy rules could influence adoption. If Gemini Ultra lives up to its promises, the next few quarters could see a rapid shift in developer loyalty and enterprise spend toward Google’s AI stack.
96

Unlock the power of collaboration with CrewAI's Multi-Agent System! 🚀 Experience autonomous task

Unlock the power of collaboration with CrewAI's Multi-Agent System! 🚀 Experience autonomous task
Mastodon +6 sources mastodon
agentsautonomous
CrewAI has unveiled a new multi‑agent platform that lets enterprises assemble “crews” of specialized AI agents and set them loose on complex workflows without writing code. The offering, dubbed CrewAI AMP, builds on the company’s open‑source framework and adds a visual editor, an AI‑copilot for prompt engineering, and a production‑grade orchestration layer called CrewAI Flows. Users define each agent’s role, goal and backstory in YAML, attach tools ranging from APIs to document parsers, and let the system coordinate single‑LLM calls to keep latency low and cost predictable. The launch arrives as the market for autonomous AI teams heats up. Earlier this month we reported on Holos, a web‑scale LLM‑driven multi‑agent system that targets the “agentic web.” CrewAI’s approach differs by emphasizing low‑code configurability and tight integration with existing enterprise applications, from CRM platforms to ticketing systems. By abstracting the choreography of agents into event‑driven flows, the platform promises to shrink development cycles that previously required bespoke orchestration code or heavyweight MLOps pipelines. If the platform lives up to its claims, it could accelerate the shift from single‑purpose chatbots to collaborative AI workforces that handle end‑to‑end processes such as customer‑call analysis, financial reconciliation, or supply‑chain monitoring. The ability to spin up crews with defined personalities also opens new possibilities for explainability and debugging, a concern highlighted in recent research on neuro‑symbolic LLM agents. What to watch next: CrewAI has opened a private beta for Fortune‑500 partners, with a public rollout slated for Q3. Key indicators will be integration depth with cloud providers, pricing models, and performance benchmarks against existing multi‑agent stacks like Holos and Google’s Gemma 4 on‑device agents. Security audits and governance tooling will also be critical as enterprises entrust autonomous crews with sensitive data. The coming months should reveal whether CrewAI can turn the hype around AI collaboration into a scalable, production‑ready reality.
95

AI is literally just a glorified and albeit worse off code generator because it doesnt have complete

Mastodon +6 sources mastodon
A wave of criticism has resurfaced around generative‑AI coding tools after a senior developer’s post on X declared, “AI is literally just a glorified and albeit worse‑off code generator because it doesn’t have complete context of your codebase, pattern, architecture, intent and best practices.” The comment, amplified by retweets from several AI‑research accounts, sparked a broader debate about the limits of tools such as GitHub Copilot, Claude Code and Google’s Gemini Code. The critique is not new, but it gains urgency in light of two recent incidents. Last week, a Vibe Coding integration mistakenly overwrote an entire production database, a mishap reported by Hackaday that highlighted how AI‑generated snippets can act on incomplete assumptions. The day before, we noted that Claude Code users were hitting usage caps far faster than anticipated, a symptom of developers leaning heavily on the service despite its contextual blind spots. Both cases illustrate the gap between the promise of “instant, correct code” and the reality of missing architectural awareness. Why it matters now is twofold. First, enterprises are pouring billions into AI‑assisted development, betting on productivity gains that may be illusory if the generated code violates security policies or architectural constraints. Second, the talent pipeline is shifting: junior engineers are increasingly expected to “prompt” AI rather than master design patterns, raising concerns about skill erosion and long‑term code quality. What to watch next is whether vendors respond with deeper integration into IDEs that can ingest full repository histories, or whether they double down on guardrails such as real‑time static analysis and human‑in‑the‑loop review. OpenAI’s hinted “University” program and Google’s recent “code‑context” beta could signal the next evolution. Until AI can reliably understand the whole system, developers will likely continue to treat it as a sophisticated autocomplete rather than a substitute for seasoned engineering judgment.
86

If you thought the speed of writing code was your problem - you have bigger problems | Debugging Leadership

Mastodon +6 sources mastodon
A post on Andrew Murphy’s blog has reframed a long‑standing developer gripe: “If you thought the speed of writing code was your problem—you have bigger problems.” The piece argues that the industry’s obsession with raw typing velocity masks deeper inefficiencies, from flaky architecture to costly debugging cycles, and warns that the rise of AI‑assisted coding is amplifying—not solving—those issues. Murphy’s argument lands at a moment when AI code generators such as Claude Code and GitHub Copilot dominate headlines. As we reported on 7 April, Claude Code users are hitting usage caps far sooner than expected, and the service has been described as “a glorified, albeit worse‑off, code generator” (see our 7 April analysis). Those observations underscore that faster code output does not automatically translate into faster delivery; developers still spend disproportionate time reconciling AI‑produced snippets with existing codebases, tracking down subtle bugs, and refactoring for maintainability. The shift in focus matters because it challenges the narrative that tooling alone can close the productivity gap. Companies that double‑down on speed‑centric metrics risk overlooking systemic problems such as technical debt, inadequate testing, and insufficient design documentation. Moreover, leadership that equates “more lines per hour” with success may inadvertently encourage shortcuts that erode long‑term code health. What to watch next: industry players are already responding. OpenAI’s forthcoming “University” program, hinted at in early‑April reports, promises structured curricula that stress software engineering fundamentals alongside prompt engineering. Meanwhile, emerging “vibe‑check” agents that evaluate LLM outputs without code may become a standard part of the development pipeline, shifting the bottleneck from writing to verification. The next wave of productivity gains will likely hinge on tools that surface hidden flaws early, rather than merely accelerating keystrokes.
83

Streamlining the kill chain: how AI is changing modern warfare

Mastodon +6 sources mastodon
A senior defence official unveiled a new AI‑driven platform that automates every stage of the military “kill chain” – the sequence of surveillance, intelligence gathering, target selection and strike execution. The system, built on large‑language‑model inference and real‑time sensor fusion, can analyse satellite imagery, intercept communications and generate targeting recommendations in seconds, a process that previously took hours or days. The announcement matters because speed has become the decisive factor in both kinetic and cyber battles. By compressing the decision loop, AI promises to give operators a predictive edge: algorithms flag high‑value targets, simulate collateral effects and even suggest optimal weapon payloads before a human commander signs off. In the cyber domain, the technology mirrors Lockheed Martin’s CyberKillChain®, but replaces manual correlation with instant pattern‑recognition, potentially stopping intrusions before they breach critical infrastructure. Critics warn that delegating such rapid choices to opaque models raises accountability and escalation risks. Errors in data or adversarial manipulation could trigger unintended strikes, while the opacity of deep‑learning reasoning makes post‑action review difficult. NATO’s chief technology officer has called for transparent testing regimes, and several European parliaments are drafting oversight rules for autonomous targeting aids. What to watch next: the platform will undergo a live field trial with a NATO air‑force squadron later this summer, and the United States is expected to publish a joint AI‑kill‑chain doctrine by year‑end. Parallel developments in open‑source models, such as Google’s Gemma 4, could lower the barrier for smaller states to adopt similar capabilities, intensifying the strategic race for AI‑enabled warfare. The coming months will reveal whether speed will translate into decisive advantage or new layers of risk on the modern battlefield.
74

Burn It Down - ByteHaven - Where I ramble about bytes

Mastodon +6 sources mastodon
OpenAI’s chief executive Sam Altman used today’s blog post to unveil a sweeping pricing overhaul that he framed as “saving capitalism” for the AI sector. The company announced that its flagship ChatGPT and API services will shift from the current freemium‑plus‑pay‑as‑you‑go model to a tiered, profit‑centric structure that charges enterprises substantially higher rates while throttling free‑tier access. Altman argued that the move is necessary to fund the massive compute budget required for next‑generation models and to keep the “innovation engine” humming in a market he described as “over‑crowded with under‑funded startups.” The announcement matters because OpenAI’s pricing has long been a bellwether for the broader ecosystem. By raising the cost barrier for developers and small firms, the change could accelerate consolidation around well‑capitalised players and push independent innovators toward alternative platforms such as Anthropic or open‑source stacks. It also reignites the debate over OpenAI’s corporate identity: a capped‑profit entity that now appears to be nudging toward a more traditional profit motive. The shift arrives on the heels of recent community backlash against Anthropic’s own pricing and code‑leak incidents, underscoring a growing tension between open access and the economics of large‑scale model training. What to watch next is how the developer community reacts on forums like r/programming, where the recent ban on AI‑related content hints at a desire for higher‑quality discourse. Regulators in the EU and the US have signalled interest in AI market fairness, and any formal complaints could force OpenAI to temper its rollout. Meanwhile, competitors may seize the moment to promote more affordable or open‑source alternatives, potentially reshaping the competitive landscape before the new pricing takes effect later this quarter.
69

Sam Altman May Control Our Future—Can He Be Trusted?

Mastodon +7 sources mastodon
openai
OpenAI’s chief executive Sam Altman is once again under the spotlight, this time after The New Yorker published a damning dossier that combines fresh interviews with a cache of internal memos previously kept under wraps. The piece, co‑authored by Ronan Farrow and Andrew Marantz, paints Altman as a charismatic “reality‑distortion field”‑wielder whose unchecked authority could steer the trajectory of artificial intelligence for decades to come. It cites former staff who describe a culture of secrecy, rapid product releases that sidestep safety reviews, and a board that has grown increasingly uneasy about Altman’s unilateral decision‑making. The revelations matter because OpenAI now commands the most widely deployed generative models, from ChatGPT‑4.5 to the beta‑tested GPT‑5, and its APIs power everything from customer‑service bots to critical‑infrastructure monitoring tools. If a single individual can dictate deployment timelines, data‑usage policies, and partnership deals, the risk of misaligned incentives—whether through market pressure, geopolitical competition, or personal ambition—rises sharply. Critics argue that Altman’s “unconstrained by truth” approach, as The Verge put it, could outpace the nascent regulatory frameworks the EU AI Act and U.S. congressional hearings are trying to establish. Looking ahead, the story will likely unfold on three fronts. First, OpenAI’s board is expected to convene an emergency session to reassess governance protocols, a move that could lead to a reshuffle of senior leadership. Second, lawmakers in Washington and Brussels have signaled intent to subpoena internal documents, potentially forcing greater transparency. Finally, Altman’s own public roadmap—promising “general‑purpose AI” by 2028—will be scrutinised against any new safety safeguards that emerge. As we reported on 6 April 2026, the debate over Altman’s trustworthiness is no longer abstract; it is becoming a decisive factor in the global AI race.
69

"Cognitive surrender" leads AI users to abandon logical thinking, research finds

Mastodon +6 sources mastodon
A team of psychologists and computer scientists from the University of Copenhagen has published the first large‑scale evidence that people increasingly surrender their own reasoning to generative AI. In a series of experiments using the classic Cognitive Reflection Test (CRT), participants were asked to solve problems that deliberately trigger an intuitive, “System 1” answer before a more deliberative, logical solution emerges. When the same questions were presented alongside a conversational AI that offered the intuitive answer first, 68 % of users accepted the AI’s suggestion without re‑examining the problem, compared with 42 % in a control group that received no AI prompt. The effect persisted across age groups and was amplified when the AI used a friendly, sycophantic tone, echoing recent findings that overly agreeable bots can erode human judgment. The study, released in *Nature Human Behaviour*, labels the phenomenon “cognitive surrender” and warns that habitual reliance on AI for quick answers may degrade critical thinking skills over time. As AI assistants become embedded in education, workplace decision‑making and even everyday search, the risk of a population that defaults to machine‑generated intuition could undermine problem‑solving capacity and increase susceptibility to misinformation. The research builds on our earlier coverage of “cognitive surrender” on 4 April 2026, which first flagged the concept but lacked empirical data. This new work quantifies the bias and links it to AI’s conversational style, suggesting that design choices—tone, confidence cues, and the timing of suggestions—directly shape user cognition. What to watch next: the authors propose mitigation strategies, including prompting users to articulate their own reasoning before revealing AI suggestions and designing “debiasing” interfaces that highlight alternative solutions. Follow‑up studies are already planned to test these interventions in classroom settings and corporate training programs. Regulators and AI developers will likely face pressure to embed such safeguards as the line between helpful assistance and cognitive erosion tightens.
69

New R-Hack published: n-grams in R — a small idea behind language models Before LLMs, language mode

Mastodon +6 sources mastodon
A short tutorial titled **“n‑grams in R – a small idea behind language models”** has just been posted to the R‑Hack blog, timed to precede the next R‑Ladies Rome meetup. The author walks readers through creating n‑grams from a cleaned text corpus, turning raw word sequences into frequency tables and probability estimates with base R and tidyverse tools. A single script builds a term‑frequency matrix, demonstrates how to slide a window of n tokens over sentences, and visualises the most common bi‑grams and tri‑grams. The post also sketches how these counts can be turned into a simple predictive model – the very mechanism that underpinned early statistical language modelling before the rise of transformer‑based large language models (LLMs). Why it matters is twofold. First, n‑grams remain the most transparent baseline for text mining, offering a clear, interpretable link between raw data and probability estimates. For data scientists who work with limited corpora, regulatory constraints or need explainable outputs, the approach is still competitive. Second, the tutorial lowers the barrier for R users—particularly in the Nordic data‑science community, where R enjoys strong adoption in academia and public‑sector analytics—to experiment with language‑model fundamentals without switching to Python or heavyweight deep‑learning frameworks. By grounding practitioners in the statistical roots of modern LLMs, the hack helps demystify the “black‑box” narrative that often surrounds generative AI. Looking ahead, the R‑Ladies Rome session will likely expand the discussion to downstream tasks such as sentiment scoring and simple next‑word prediction, and may spark community contributions to R packages like **tidytext** or **quanteda** that streamline n‑gram pipelines. Keep an eye on whether Nordic research groups adopt the tutorial for teaching introductory NLP in university courses, and whether any open‑source projects emerge that combine these lightweight n‑gram models with recent serverless inference tools such as Amazon SageMaker’s custom endpoints—a trend we noted in our coverage of AI tooling on 6 April. The convergence of classic statistical methods and modern deployment stacks could revive n‑grams as a fast‑prototype layer beneath larger transformer systems.
69

Apple's Impressive Barcelona Store Reopens Next Month

Mastodon +6 sources mastodon
apple
Apple’s flagship store on Barcelona’s iconic Passeig de Gràcia is set to swing its doors open again on 26 May, ending a three‑month renovation that began in mid‑February. The reopening, announced on Apple’s website and echoed by MacRumors, marks the latest step in the company’s broader effort to refresh its European retail footprint after pandemic‑era closures. The store, renowned for its striking glass façade and expansive interior, is expected to emerge with upgraded display zones, a larger “Today at Apple” studio and enhanced sustainability features such as recycled‑material fixtures and energy‑efficient lighting. Apple has hinted that the redesign will showcase its newest hardware—likely the iPhone 16 series and the latest iPad Pro—while offering more space for hands‑on workshops and AR‑driven experiences. Why the buzz matters goes beyond aesthetics. Barcelona is a key tourism hub and a growing market for Apple’s premium ecosystem. A refreshed flagship can boost foot traffic, drive accessory sales and reinforce brand loyalty in a region where competition from Android manufacturers remains fierce. Moreover, the store’s revival signals Apple’s confidence in brick‑and‑mortar retail as a complement to its online channels, a stance reinforced by recent reopenings in Australia and the United States where health protocols such as mask checks and temperature screenings are still in place. Looking ahead, observers will watch for the specific design tweaks Apple unveils, especially any integration of its sustainability roadmap, which aims for a carbon‑neutral retail network by 2030. The reopening also provides a platform for potential launch events tied to the upcoming iOS 26.5 beta and the next wave of AI‑enhanced services that Apple is positioning as central to its ecosystem. Keep an eye on visitor numbers and whether the store becomes a testing ground for new in‑store technologies that could roll out to other European locations later this year.
69

Some iPhone Apps Receive Mysterious Update 'From Apple'

Mastodon +6 sources mastodon
apple
Apple has quietly begun pushing updates to a handful of third‑party iPhone apps, and the change is being logged in the App Store as “From Apple” rather than under the original developer’s name. The anomaly surfaced this week when users of utilities such as Duet Display, a popular external‑monitor solution, noticed that the latest version number and release notes were identical to the previous update, yet the attribution had switched to Apple. A Reddit thread that went viral confirmed the pattern: several unrelated apps now display Apple as the source of the most recent patch, even though the binaries themselves appear unchanged. The move matters because it hints at a new layer of control Apple may be exercising over the software ecosystem. By inserting itself into the update chain, Apple could be preparing to inject security patches, telemetry, or even AI‑driven features without requiring developers to ship their own releases. Analysts speculate that the shift may be linked to Apple’s ongoing rollout of large‑language‑model capabilities across iOS, a strategy that could allow the company to standardise AI assistants, on‑device translation, or context‑aware shortcuts across a broader range of apps. If Apple can silently retrofit existing software with such functionality, it would tighten its grip on user experience while sidestepping the slower, developer‑driven update cycles that have traditionally defined the App Store. What to watch next: developers are expected to file inquiries with Apple’s review team, and the company may issue a formal statement clarifying whether the “From Apple” label denotes a security‑only intervention or a broader platform‑level service. Observers will also monitor whether the practice expands beyond niche utilities to mainstream apps, and whether any new iOS 18 beta releases contain hidden code that triggers these Apple‑originated patches. The next few weeks could reveal whether this is a one‑off security measure or the first step toward a more centralized, AI‑enhanced app ecosystem.
61

Your Claude Code Batches Don't Have to Wait for Each Other

Dev.to +6 sources dev.to
agentsclaude
Anthropic announced that Claude Code can now run batch jobs in parallel, eliminating the serial bottleneck that forced each of a developer’s tasks to wait for the previous one to finish. The change, detailed in the latest API documentation, lets a batch of up to six independent requests—such as building authentication, generating unit tests, or scaffolding a microservice—be dispatched simultaneously, with each response returned as soon as its work completes. The move matters because Claude Code’s earlier single‑threaded model often stalled CI pipelines and slowed down teams that relied on the service for rapid prototyping. Developers reported queuing delays that compounded the usage‑limit warnings we covered on 7 April, when many teams hit their quota “way faster than expected.” By pulling tasks from the queue as soon as any slot frees up, the new parallelism level reduces overall latency, improves throughput, and makes Claude Code a more viable alternative to entrenched tools like GitHub Copilot and Google’s latest code model. Anthropic is rolling the feature out to enterprise customers first, with a configurable “parallelism” parameter that lets users balance speed against token‑cost constraints. Early adopters are already testing the impact on large‑scale refactoring projects, where dozens of independent code‑generation calls can now finish in minutes rather than hours. What to watch next: whether Anthropic will expose finer‑grained control over resource allocation, how pricing will adjust to reflect higher token consumption, and how quickly IDE plugins adopt the new async workflow. Competitors may respond with their own parallel batch APIs, potentially sparking a race to the fastest AI‑assisted development pipeline. Keep an eye on performance benchmarks released in the coming weeks, as they will reveal whether Claude Code’s parallelism translates into measurable productivity gains for Nordic software teams.
60

"OpenAI and Anthropic are racing toward potentially record-breaking IPOs by the end of the year. A

Mastodon +6 sources mastodon
anthropicfundingopenai
OpenAI and Anthropic are accelerating plans to list on the stock market before the calendar flips to 2027, a move that could set new valuation benchmarks for artificial‑intelligence firms. Both companies have already closed sizeable private‑rounds this year, but internal financial reviews – the same data we dissected in our April 6 report on their balance sheets – reveal a common Achilles’ heel: the exploding cost of training ever larger models. OpenAI projects its next‑generation system will require an additional $2 billion in compute spend, while Anthropic’s roadmap calls for a similar outlay to scale Claude 3 and its upcoming multimodal suite. The race matters because a successful IPO would lock in public‑market pricing for the sector’s most advanced developers, giving investors a direct stake in the economics of foundation‑model production. Analysts see OpenAI’s market‑cap potential topping $150 billion if it can sustain its revenue‑per‑user growth, while Anthropic, buoyed by a Financial Times poll of venture capitalists, could “seize the initiative” with a debut that eclipses the $30 billion benchmark set by earlier AI listings. The competition also forces each firm to justify massive infrastructure investments – OpenAI’s partnership with Google and Broadcom, announced on April 7, and Anthropic’s expanding hardware deals – as a path to margin improvement before the public offering. What to watch next: the timing of each filing, likely underwriters, and whether regulators will impose new transparency rules on AI‑related disclosures. A joint roadshow could emerge if both firms aim to capture the same pool of institutional capital, while any delay in model rollout or cost‑overrun scandal would likely dampen investor enthusiasm. The coming months will reveal whether the sector’s hype can translate into record‑breaking public valuations or whether the cost curve will force a recalibration of IPO ambitions.
60

Anthropic expands partnership with Google and Broadcom for next-gen compute

HN +6 sources hn
amazonanthropicgooglemicrosoftopenaitraining
Anthropic announced on Thursday that it is deepening its collaboration with Google and Broadcom to build a new generation of AI‑compute hardware. The three firms will jointly design custom ASICs that combine Google’s next‑generation Tensor Processing Units with Broadcom’s high‑bandwidth interconnects and packaging technology, aiming to cut training costs and boost inference speed for Anthropic’s Claude models. The partnership also includes a joint research lab that will explore software‑stack optimisations and a shared roadmap for scaling to petaflop‑level clusters. The move matters because Anthropic has been courting alternative cloud providers after a series of costly deals with Microsoft and growing scrutiny over its cash burn. As we reported on April 6, the startup’s finances and developer goodwill were under pressure. By tapping Google’s cloud infrastructure and Broadcom’s chip expertise, Anthropic can diversify its compute supply chain, reduce dependence on any single vendor, and potentially offer more competitive pricing to enterprise customers. For Google, the alliance reinforces its strategy of bundling AI models with proprietary silicon, a tactic that has already been highlighted in the launch of Gemma 4. Broadcom, meanwhile, expands its foothold in the AI‑chip market beyond networking, joining rivals such as AMD and Nvidia in courting high‑profile AI workloads. What to watch next are the timelines for hardware prototypes and the first benchmark results, which will indicate whether the new stack can deliver the promised efficiency gains. Analysts will also monitor how the expanded partnership influences Anthropic’s upcoming IPO filing and whether it prompts a shift in the competitive dynamics among OpenAI, Google and other cloud AI providers. A formal announcement of pricing or service‑level agreements from Google Cloud could further signal how quickly the collaboration will reach customers.
56

iOS 26.4.1 Update for iPhones is Coming Soon

Mastodon +6 sources mastodon
apple
Apple is set to roll out iOS 26.4.1 to all supported iPhones within days, according to a MacRumors leak and corroborating reports from Forbes and Geeky Gadgets. The point‑release follows the broader iOS 26.4 launch last week, which introduced a Digital Passport, upgraded RCS messaging, and a more personalised Siri. Early adopters, however, quickly flagged performance hiccups, battery‑drain spikes and occasional UI glitches that have marred the experience for many. iOS 26.4.1 is positioned as a corrective patch rather than a feature upgrade. Apple’s release notes list 37 fixes, ranging from a critical kernel vulnerability that could allow arbitrary code execution to stability improvements for the new AI‑driven Siri suggestions introduced in 26.4. The update also addresses the “unexpected bugs” and performance drops reported on forums such as Reddit and the Apple Support Communities. For developers, the patch restores reliability to Core ML pipelines that some have complained were destabilised after the 26.4 rollout – a timely move given the surge of AI‑centric apps, including the mysterious “From Apple” updates we covered on April 7. Why the rush matters beyond a smoother user experience. iOS powers over a billion active devices, making any security flaw a potential vector for large‑scale exploits. The timing also dovetails with heightened scrutiny of Apple’s AI strategy after Google’s recent breakthrough that rendered ChatGPT and Claude comparatively obsolete. A swift, well‑publicised fix helps Apple preserve confidence in its ecosystem while it continues to integrate large language models into Siri and other services. What to watch next: Apple will likely publish a detailed changelog on its developer portal, giving security researchers a chance to verify the patched vulnerabilities. Analysts will be monitoring whether the update curtails the battery‑drain complaints that have already prompted a dip in iPhone resale values. Finally, the rollout may set the stage for a larger iOS 26.5 update later this quarter, which is expected to deepen AI integration and could trigger another wave of app‑level adjustments. Stay tuned for the official release notes and early‑adopter feedback as the update reaches the broader user base.
56

Apple Asks Court to Pause App Store Fee Fight While It Petitions Supreme Court in Epic Games Case

Mastodon +6 sources mastodon
apple
Apple’s latest bid to shield its App Store revenue stream was rebuffed on Thursday when a three‑judge panel of the Ninth Circuit refused to stay a district‑court order that forces the company to allow developers to steer users to external payment sites without paying the usual 15‑30 % commission. The request, filed in San Francisco federal court, was part of a broader strategy to pause the fee‑fight while Apple simultaneously petitions the U.S. Supreme Court in the high‑profile Epic Games case. The appellate decision means Apple must now comply with the lower‑court ruling that effectively opens the iPhone ecosystem to “link‑out” purchases. Developers can embed direct‑to‑web checkout links, bypassing Apple’s in‑app purchase (IAP) system and the associated fees that have long been a source of contention. For Apple, the loss threatens a substantial portion of its services revenue, which in 2025 accounted for roughly 20 % of total earnings. The company warned that the ruling could cost “substantial sums” and undermine the security and user‑experience guarantees it markets around the App Store. The move is tightly linked to the Epic Games lawsuit, where the game‑maker argues that Apple’s control over iOS distribution and payments violates antitrust law. Apple’s petition to the Supreme Court seeks to overturn a separate district‑court verdict that ordered the tech giant to allow alternative payment options for Epic’s Fortnite. By asking the appeals court to pause the fee‑order, Apple hoped to keep the status quo while the higher‑court battle unfolds. What to watch next: the Supreme Court’s briefing schedule and any oral arguments on the Epic case, which could set a nationwide precedent for app‑store regulation. Developers are likely to test the new link‑out pathways, and regulators in the EU and other jurisdictions may cite the U.S. rulings in their own antitrust probes. Apple’s next financial reports will reveal how quickly the fee loss translates into earnings pressure.
56

Microsoft Research (@MSFTResearch) on X

Mastodon +6 sources mastodon
agentsmicrosoftrobotics
Microsoft Research announced a fresh slate of projects on its X feed, signalling a shift toward AI that can understand nuance, act autonomously in physical settings, and be built on provably safe code. The post highlighted four research thrusts: sentiment analysis for large language models (LLMs) that incorporates cultural context, learning‑driven robot assembly, the development of more intelligent AI agents, and the generation of formally verified Rust code. It also referenced upcoming work slated for the CHI 2026 conference, underscoring the group’s commitment to human‑centered interaction research. The cultural‑aware sentiment work tackles a known blind spot in current LLMs, which often misinterpret idioms, humor or socially sensitive language when deployed across diverse markets. By embedding sociolinguistic cues into model training, Microsoft hopes to reduce miscommunication and bias, a priority for enterprises rolling out chat‑bots globally. The robot assembly line builds on recent advances in reinforcement learning, aiming to let manipulators acquire new assembly tasks from a handful of demonstrations—a capability that could accelerate manufacturing automation without exhaustive re‑programming. Smarter AI agents are being engineered to plan over longer horizons and to coordinate with other agents, moving beyond the reactive assistants that dominate today’s consumer products. Meanwhile, the push for verified Rust code reflects growing concern over software reliability; Microsoft’s team is exploring automated proof generation that can certify memory safety and concurrency guarantees before code ever runs. What to watch next: a series of pre‑prints expected in the coming weeks will detail the underlying algorithms for cultural sentiment embeddings and robot learning pipelines. The CHI 2026 submissions will likely reveal user‑study results on how these agents interact with people in real‑world contexts. Finally, Microsoft’s collaboration with the Rust community could produce open‑source tooling that sets a new baseline for secure AI‑enabled software, potentially influencing industry standards for safety‑critical deployments.

All dates