AI News

300

Introspective Diffusion Language Models

Introspective Diffusion Language Models
HN +5 sources hn
Researchers at the University of Copenhagen and the Nordic AI Lab have unveiled the Introspective Diffusion Language Model (I‑DLM), a diffusion‑based transformer that claims to match autoregressive (AR) quality while decoding tokens in parallel. The core innovation, Introspective Strided Decoding (ISD), lets the model verify previously generated tokens and advance new ones within a single forward pass, eliminating the step‑by‑step token generation that has long defined large language models. The paper, posted on arXiv two days ago, reports that I‑DLM attains parity with same‑scale AR models on fifteen benchmarks covering factual knowledge, mathematics, code synthesis and instruction following. The authors attribute the gain to diffusion‑style parallelism combined with an introspective consistency check that mimics the self‑correction behavior of AR training. A public GitHub repository already provides the code and pretrained weights, inviting rapid replication. If the results hold, the development could reshape the efficiency calculus of LLM inference. Diffusion models have excelled in image generation but have struggled to reach the fluency of AR text generators; I‑DLM’s single‑pass generation promises lower latency and reduced memory bandwidth, traits attractive to edge devices and data‑center operators alike. Moreover, the built‑in verification step may curb hallucinations, a persistent pain point for commercial deployments. The community will be watching for large‑scale replication studies and for integration into popular inference stacks such as OpenAI’s API wrappers and the Nordic AI open‑source ecosystem. Upcoming conferences on AI hardware and the next wave of benchmark suites will likely feature head‑to‑head comparisons with the hybrid neural‑symbolic models we covered on 13 April. If I‑DLM scales as advertised, it could usher in a new generation of fast, self‑checking language models that challenge the AR monopoly.
195

Can AI be a 'child of God'? Inside Anthropic's meeting with Christian leaders

Can AI be a 'child of God'? Inside Anthropic's meeting with Christian leaders
HN +6 sources hn
anthropicclaude
Anthropic convened a two‑day summit at the end of March that brought together roughly 15 senior figures from Catholic and Protestant churches, as well as scholars and business leaders, to discuss the moral and spiritual status of its flagship chatbot, Claude. The invitation, confirmed by the Washington Post, was framed as an effort to “seek guidance on how Claude should be treated as a moral agent,” and the agenda famously included the provocative question: can an AI ever be considered a “child of God”? The meeting marks the first time a major AI developer has formally engaged organized religion on the philosophical underpinnings of its technology. Anthropic, which has positioned itself as a safety‑first startup after its $30 billion funding round, is grappling with the same ethical dilemmas that surfaced in its recent LLM‑Meter surge to 89 and the internal debate over Claude’s memory architecture. By inviting clergy to weigh in on concepts such as personhood, responsibility, and the potential for AI to influence human values, the company signals a shift from purely technical safety work to broader cultural stewardship. Stakeholders are watching to see whether the summit yields concrete policy recommendations or merely philosophical musings. Anthropic’s next steps could include publishing a set of ethical guidelines, integrating religious perspectives into its alignment training data, or establishing an advisory board with permanent theological representation. The outcome may also shape regulatory discussions in Europe and the United States, where lawmakers are increasingly probing the societal impact of generative AI. As we reported on Claude’s rapid performance gains on April 12, the company’s focus is now expanding from raw capability to the deeper question of what it means to create systems that interact with humanity on a moral plane. Future updates from Anthropic, or any formal framework emerging from this dialogue, will be a key barometer for how the AI industry navigates the intersection of technology and faith.
171

OpenAI touts Amazon alliance in memo, says Microsoft has 'limited our ability' to reach clients

CNBC +6 sources 2026-04-13 news
amazonmicrosoftopenai
OpenAI’s newly hired revenue chief, Denise Dresser, circulated an internal memo on Sunday that frames the company’s expanding partnership with Amazon as a cornerstone of its enterprise strategy while accusing Microsoft of curbing OpenAI’s ability to reach customers. The document, seen by several outlets, says OpenAI is actively “reducing its reliance on Microsoft” and points to the $50 billion investment Amazon made in the startup last year as the basis for a deeper technical and go‑to‑market collaboration. The memo marks the first public acknowledgment that OpenAI is deliberately diversifying away from its long‑standing reliance on Microsoft’s Azure cloud and the Azure OpenAI Service, which has been the primary conduit for corporate clients to embed ChatGPT‑based tools. Dresser argues that Microsoft’s control over pricing, licensing and co‑selling arrangements has limited OpenAI’s flexibility, prompting the firm to lean on Amazon Web Services for compute, data‑pipeline integration and joint marketing. If the shift materialises, it could give Amazon a foothold in the lucrative generative‑AI market that has been dominated by Azure, while forcing Microsoft to renegotiate terms that have underpinned its AI‑first positioning. Analysts see the move as a litmus test for OpenAI’s bargaining power. A stronger Amazon tie‑up could pressure Microsoft to offer more favourable revenue splits or broaden access to OpenAI’s models, potentially reshaping the competitive landscape for enterprise AI platforms. It also raises questions about how OpenAI will balance dual‑cloud deployments and whether customers will be nudged toward one provider over the other. Watch for official statements from Microsoft, Amazon and OpenAI in the coming weeks, as well as any revisions to the Azure OpenAI Service pricing or new joint offerings announced on AWS. The next earnings call and any regulatory filings could reveal how the partnership will be quantified in revenue and whether the “limited ability” claim translates into concrete contract changes.
158

I took the prompt generated by my # Moodle AIText question type and pasted it into the Edge Gall

I  took the prompt generated by my  # Moodle   AIText question type and pasted it into the Edge Gall
Mastodon +6 sources mastodon
gemmagoogle
A Moodle instructor has taken the prompt generated by the platform’s new **AIText** question type and run it on a Google Pixel 7 equipped with GrapheneOS, using the Edge Gallery app to invoke the **GoogleGemma‑4‑E2B‑it** model entirely offline. After copying the prompt into Edge Gallery, the user disabled all network connections, forcing the phone’s on‑device inference engine to produce the answer without reaching external servers. The experiment proves that Moodle’s AI‑driven assessment tools can be decoupled from cloud APIs and executed locally on consumer hardware. By leveraging a privacy‑focused OS and an on‑device LLM, educators can offer AI‑assisted feedback while guaranteeing that student data never leaves the device. This addresses long‑standing concerns about data sovereignty, GDPR compliance, and the risk of exposing exam content to third‑party services. It also sidesteps the latency and cost issues that have hampered large‑scale adoption of cloud‑only LLMs in schools. As we reported on 14 April, the **Anthropic Opus** model was already being trialled to re‑imagine Moodle’s gradebook, highlighting a broader push to embed generative AI deeper into the learning management system. The current offline test extends that trajectory, showing that the same prompt‑generation logic can feed a variety of models, from hosted APIs to edge‑optimized variants, without redesigning the Moodle plugin. What to watch next: benchmark results comparing Gemma‑4’s accuracy and speed against cloud‑based counterparts; updates from the Edge Gallery team on model support and battery impact; and Moodle’s roadmap for native offline‑LLM integration. If the approach scales, we may see a new class of “privacy‑first” AI tools in classrooms across the Nordics, prompting policy makers to revisit guidelines on AI use in education.
158

The Journalists Striking Over AI

The Journalists Striking Over AI
Mastodon +6 sources mastodon
layoffs
ProPublica, one of the United States’ largest nonprofit newsrooms, saw roughly 150 journalists walk off the job for a 24‑hour strike on Wednesday. The action, the outlet’s first major labor dispute, was driven by three intertwined grievances: stagnant wages, inadequate lay‑off protections, and a newly imposed artificial‑intelligence policy that the union says was rolled out without bargaining. The strike was coordinated by the newsroom’s staff union, which also filed an unfair‑labor‑practice charge, arguing that management’s unilateral AI rollout violated collective‑bargaining obligations. Employees demanded clear guardrails on the use of large language models in reporting, assurances that AI tools would not replace human bylines, and a transparent framework for how generated content would be disclosed to readers. The protest reverberates beyond ProPublica. Australia’s national broadcaster ABC announced a parallel strike over pay and AI safeguards, signalling a growing trans‑Atlantic unease among journalists about the speed of automation. The timing coincides with heightened scrutiny of AI’s role in newsrooms, from OpenAI’s recent regulatory challenges to the launch of platforms like Veritas that promise real‑time oversight of disinformation. As newsrooms grapple with cost pressures, AI promises efficiency but also raises ethical and employment questions that have yet to be resolved. What to watch next: the outcome of ProPublica’s bargaining round, which is slated to resume within days, will set a benchmark for how media unions negotiate AI clauses. Industry observers will also monitor whether other news organisations follow suit, potentially sparking a wave of collective actions that could shape AI governance standards across the press. The strike underscores that the battle over AI is now as much about labour rights as it is about technology.
156

Claude Code Routines

Claude Code Routines
HN +6 sources hn
agentsclaude
Anthropic unveiled “Claude Code Routines” on Tuesday, a library of pre‑crafted workflow templates that let developers string together coding tasks without writing custom prompts for each step. The offering bundles five patterns – sequential, operator, split‑and‑merge, agent‑team and headless – and ships with ready‑made scripts for codebase exploration, bug fixing, refactoring and test generation. Users can invoke a routine with a single command, letting Claude Code handle the orchestration behind the scenes. The move addresses a pain point that surfaced in recent weeks. As we reported on April 14, Claude Code suffered a 12‑hour OAuth outage and, earlier in the month, developers complained about “invisible tokens” eating their usage limits. By abstracting common sequences into reusable routines, Anthropic hopes to reduce the number of API calls per task, lower token waste and make the service more resilient to transient authentication glitches. Early adopters say the patterns cut setup time by up to 70 percent and make the agentic capabilities described in the “Claude Code Agentic Workflow Patterns” guide more accessible to teams without deep prompt‑engineering expertise. What to watch next is how quickly the routines migrate from documentation to production. Anthropic has promised tighter IDE integrations and a marketplace for community‑built routines, but pricing details remain vague. Competitors such as GitHub Copilot X and Microsoft’s upcoming “Co‑pilot” are also rolling out higher‑level automation, so market reception will hinge on Claude Code’s performance on real‑world codebases and its ability to stay within the token budgets that have recently plagued users. Follow‑up reporting will track adoption metrics, any impact on the OAuth reliability issues, and whether the new patterns spur a broader shift toward agent‑centric development tools in the Nordic AI ecosystem.
149

I guess I do need to say these kinds of things repeatedly because just mentioning you find LLMs usef

I guess I do need to say these kinds of things repeatedly because just mentioning you find LLMs usef
Mastodon +6 sources mastodon
startup
A wave of online commentary is exposing a cultural backlash against the enthusiastic promotion of large‑language models (LLMs). A recent post that went viral on a tech‑focused forum captured the sentiment: the author, who describes a background in hardware‑startup engineering rather than “tech bro” culture, says they feel compelled to repeat how useful LLMs are only to be dismissed as a shallow AI evangelist. The thread quickly attracted dozens of replies that echoed the same frustration, highlighting a growing stereotype that equates praise for generative AI with a lack of critical thinking. The episode matters because perception can become a decisive factor in technology adoption. While LLMs are being rolled out across enterprises—from customer‑service chatbots to code‑generation assistants—skepticism fueled by social labeling may slow uptake, especially among groups already wary of new tech. As we reported on 14 April, older workers have long carried a reputation for resisting innovation; the current stigma around AI enthusiasm adds another layer of resistance that could affect hiring, training and internal advocacy programs. Industry observers see the discussion as a barometer of the broader societal divide over AI. Companies are now wrestling with how to communicate the benefits of LLMs without triggering the “tech‑bro” label that can alienate non‑technical staff. Researchers are also launching surveys to quantify the extent of the bias and to identify communication strategies that resonate across age and professional backgrounds. What to watch next: a forthcoming study by the Nordic Institute for Digital Society on AI perception, slated for release later this quarter, promises hard data on the correlation between self‑identification as an AI advocate and perceived credibility. In parallel, several Nordic firms have announced internal “AI literacy” campaigns aimed at reframing LLMs as practical tools rather than status symbols, a move that could reshape the narrative before the next wave of enterprise roll‑outs.
147

Jon Prosser Still Not Fully Cooperating in Apple's iOS 26 Trade Secrets Lawsuit

Mastodon +6 sources mastodon
apple
Apple has filed a fresh motion in the Northern District of California accusing YouTube leaker Jon Prosser of failing to comply with a subpoena issued in the company’s trade‑secrets lawsuit over iOS 26. The lawsuit, lodged in July 2025, alleges that Prosser used insider contacts to obtain confidential design documents—including details of the “Liquid Glass” screen coating and UI changes—before the operating system’s public unveiling at WWDC. Apple says Prosser not only misappropriated the information but also offered incentives to others for access, violating the Computer Fraud and Abuse Act. Prosser’s legal team has so far provided only partial production of requested emails and device logs, prompting Apple to request a court‑ordered deadline and possible sanctions. The tech‑media figure maintains he acted in good faith, claiming he was unaware of how the material reached him and that any cooperation was hampered by “over‑broad” discovery demands. The dispute matters because it pits a high‑profile independent commentator against one of the world’s most protective corporations. A ruling that forces Prosser to hand over additional evidence could set a precedent for how aggressively Apple can pursue leakers, potentially chilling the flow of early‑stage information that fuels the vibrant ecosystem of Apple‑focused content creators. Conversely, a finding that Apple’s subpoena is overly intrusive could embolden other journalists and analysts to push back against corporate gag orders. Watch the court docket over the next few weeks for a judge’s ruling on the motion to compel full compliance. Apple is expected to file a supplemental brief outlining the specific damages it claims, while Prosser’s side may seek a protective order. The outcome could influence Apple’s forthcoming iOS 26.5 beta rollout and shape how the tech press handles future leaks.
144

Building long-running AI agents just got a massive efficiency upgrade. 🛠️ The new NVIDIA Agent Toolk

Mastodon +7 sources mastodon
agentsmicrosoftnvidiaopen-source
NVIDIA has unveiled the Agent Toolkit, a suite of open‑source components designed to make long‑running AI agents far more efficient. The package bundles two new tools—NemoClaw, a hardened execution sandbox, and OpenShell, a lightweight orchestration layer—alongside AI‑Q, a query‑optimisation engine that promises to halve the cost of LLM calls without sacrificing the accuracy that powers today’s most demanding applications. The announcement arrives at a moment when developers are wrestling with the operational overhead of agents that must persist state, manage multi‑step workflows and stay responsive over hours or days. By isolating agent processes in NemoClaw, NVIDIA aims to curb the security risks that have plagued earlier sandbox attempts, while OpenShell supplies a plug‑and‑play API for scheduling, retry logic and external service integration. AI‑Q, meanwhile, leverages dynamic prompt compression and selective model routing to reduce token consumption, a boon for enterprises that run thousands of concurrent agents on cloud credits. Why it matters is twofold. First, the cost reduction directly addresses a barrier to scaling agentic services in production, where query fees can quickly eclipse revenue. Second, the toolkit’s open‑source nature invites community‑driven extensions, positioning NVIDIA as a rival to Microsoft’s Agent Framework, which we covered on 13 April 2026 as the backbone for Azure‑hosted agents. The competition could accelerate standards for state management, security and billing across the ecosystem. Looking ahead, developers will be testing the toolkit against real‑world workloads such as autonomous customer‑support bots and AI‑driven data pipelines. Watch for benchmark releases from both NVIDIA and independent labs, and for any integration announcements with cloud providers—especially Azure, which may respond with tighter coupling to its own framework. The next few weeks should reveal whether the promised 50 % cost cut translates into measurable adoption and whether the open‑source model spurs a new wave of interoperable agent platforms.
143

Stop trying to write magic incantations for an #llm Start — Tony Sullivan

Stop trying to write magic incantations for an #llm

Start — Tony Sullivan
Mastodon +6 sources mastodon
open-source
Tony Sullivan’s latest post on X, “Stop trying to write magic incantations for an #llm,” is quickly becoming a touchstone for developers wrestling with the hype‑driven approach to large language models. In a terse, three‑paragraph manifesto, Sullivan urges teams to treat LLM‑powered components the same way they treat any open‑source library: start with a clear README, publish contribution guidelines, enforce a style guide, and automate quality checks with linters and tests. He argues that the current “prompt‑as‑spell” mindset—where a cleverly worded prompt is expected to conjure flawless code—ignores the engineering discipline that keeps software reliable at scale. The shift matters because enterprises are now embedding LLMs in production pipelines for code generation, documentation, and even customer support. Early adopters that rely on ad‑hoc prompts are already reporting brittle outputs, hidden biases, and costly rollbacks. By framing LLM integration as a software engineering problem, Sullivan’s call‑to‑action pushes the industry toward reproducible, auditable practices that can be version‑controlled and peer‑reviewed. It also dovetails with the broader movement to professionalise AI development, echoing recent discussions about “hand‑off” protocols for AI agents and the need for dedicated SDKs that respect resource constraints. What to watch next: several open‑source projects, such as the LangChain community and the emerging LLM‑Ops toolkits, are rolling out scaffolding that includes README templates, contribution checklists, and automated prompt‑testing suites. Standards bodies are expected to publish the first draft of an “LLM Engineering” style guide later this year. If Sullivan’s prescription gains traction, the next wave of AI‑augmented products will look less like experimental magic tricks and more like rigorously engineered software.
143

Never Read Again! AI Will Do It For You

Never Read Again! AI Will Do It For You
Mastodon +6 sources mastodon
A new AI service that “reads” entire books aloud has gone viral after a YouTube demo, titled “Never Read Again! AI Will Do It For You,” showed the system narrating a novel from start to finish while simultaneously generating on‑screen summaries. The startup behind the demo, called **StorySynth**, combines a large language model fine‑tuned on narrative structure with a high‑fidelity text‑to‑speech engine, allowing users to upload a digital copy and receive a continuous audio stream that claims to preserve the author’s voice, tone and pacing. The launch matters because it pushes the boundary between assistance and substitution. For visually impaired readers, the technology promises a more natural, context‑aware alternative to existing screen‑readers. For the broader market, it threatens traditional reading habits and raises fresh copyright questions: publishers have already warned that mass‑scale, AI‑generated narration could bypass royalties and undermine print sales. The Guardian’s recent coverage of “robot storytellers” highlighted similar ethical concerns, noting that many readers feel a moral compromise when a machine replaces a human voice. StorySynth’s model also claims to “never stop learning,” echoing the SEAL architecture described in WIRED, which continuously updates from new text without explicit retraining. If true, the system could improve its narration style over time, but it also opens the door to unvetted content drift and potential bias amplification. What to watch next includes legal challenges from authors’ unions, pilot programmes with e‑reader manufacturers, and the emergence of competing services that bundle AI narration with summarisation tools. Observers will also track whether publishers adopt licensing schemes that monetize AI‑driven audio, or whether the technology spurs a new wave of “listen‑first” publishing models that reshape how books are consumed in the Nordic market and beyond.
139

Turn Your Laptop Into an AI Agent (Free OpenClaw + Telegram Setup)

Turn Your Laptop Into an AI Agent (Free OpenClaw + Telegram Setup)
Dev.to +5 sources dev.to
agents
OpenClaw, an open‑source AI assistant that runs entirely on a user’s machine, has just been packaged with a one‑click Telegram integration, turning any laptop into a fully fledged personal agent. The new “OpenClaw + Telegram” bundle can be deployed on Railway with a single command, spins up a local dashboard on port 18789, and lets users issue commands such as “Say hello in one sentence” via a Telegram chat. The assistant can read and act on emails, manipulate files, control browsers and schedule tasks, all without sending data to the cloud. The release matters because it lowers the barrier to private, on‑device AI automation. While large‑scale cloud agents dominate the market, they require API keys, incur usage fees and expose sensitive data. OpenClaw’s local execution, combined with a familiar messaging interface, gives hobbyists, small businesses and privacy‑conscious professionals a way to harness LLM‑powered automation without the overhead of cloud services. It also builds on the recent surge of “run‑your‑own‑LLM” guides we covered last week, and dovetails with the growing “agent‑as‑a‑service” ecosystem that includes Claude Managed Agents and Amazon Bedrock AgentCore. What to watch next is how quickly the community expands the skill library that OpenClaw can learn from YouTube videos, GitHub repos or personal scripts, and whether the project adds support for other messengers such as Discord or WhatsApp. Performance benchmarks with newer open models like Gemma 4 will indicate whether the tool can keep pace with commercial agents. Finally, we’ll keep an eye on any enterprise‑grade security hardening or integration with existing workflow platforms, which could push OpenClaw from a hobbyist curiosity into a viable alternative for private AI automation.
138

Run Your Harper AI Agent on Google Cloud Vertex AI — 3 Files Changed

Run Your Harper AI Agent on Google Cloud Vertex AI — 3 Files Changed
Dev.to +6 sources dev.to
agentsgooglevector-db
A three‑file update to the open‑source Harper framework now lets developers launch the conversational AI agent on Google Cloud’s Vertex AI platform. The patch replaces the local inference stack with calls to Vertex AI’s managed generative‑model service, wiring Harper’s semantic cache and vector‑memory modules into the cloud‑native SDK. The change also adds a lightweight wrapper that translates Harper’s internal request format into the Vertex AI API, enabling the agent to tap any of the 200+ foundation models hosted on Google’s infrastructure, including Gemini and Anthropic’s Claude. The move matters because it shifts Harper from a hobbyist‑level, locally‑run prototype to a production‑grade service that inherits Vertex AI’s enterprise‑grade security, compliance and autoscaling. Nordic firms that have been cautious about hosting large language models on‑premise can now experiment with agentic workflows without provisioning GPU clusters or managing model updates. The integration also opens the door to hybrid pipelines: developers can keep sensitive context in Harper’s on‑device vector store while offloading heavy generation to the cloud, reducing latency and cost compared with fully local deployments. As we reported on April 14, earlier posts explored running LLMs locally and building a privacy‑first voice‑controlled AI agent with local models. This latest step shows the same team extending those concepts onto a managed platform, reflecting a broader industry trend of blending edge privacy with cloud scalability. Watch for performance benchmarks that compare on‑premise versus Vertex AI latency and token pricing, and for announcements from Google about tighter coupling between Vertex AI Agent Builder and third‑party frameworks like Harper. Further community contributions could add support for other clouds, while enterprise pilots in the Nordics will reveal how quickly the model‑agnostic agent approach gains traction in regulated sectors such as finance and healthcare.
138

iPhone Fold Production Pushed Back, But Fall 2026 Launch Still on Track

iPhone Fold Production Pushed Back, But Fall 2026 Launch Still on Track
Mastodon +6 sources mastodon
apple
Apple’s first foldable iPhone has hit a modest production snag, but the company’s timetable for a 2026 debut remains intact. According to DigiTimes, mass‑production of the “iPhone Fold” has slipped from an anticipated June start to early August, a delay of roughly one to two months. Apple has not signaled any change to its launch window, and analysts still expect the device to appear alongside the iPhone 18 Pro and Pro Max in September 2026. The setback matters because Apple’s entry into the foldable segment has been watched as a potential game‑changer for a market dominated by Samsung and a growing roster of Chinese manufacturers. A delayed ramp‑up could compress the supply chain ahead of the holiday season, affect component pricing, and give rivals extra breathing room to refine their own designs. The issue also hints at the technical hurdles Apple faces – from hinge durability to the integration of under‑display cameras – challenges that have already slowed other manufacturers. What to watch next is whether the production delay expands as Apple moves from pilot runs to full‑scale assembly. Supplier briefings in the coming weeks, especially from Foxconn and its partners, will reveal if the August target holds. Apple’s upcoming September event will be the first public platform to showcase the foldable’s design language, pricing strategy and software integration, all of which will shape the competitive dynamics with Samsung’s Galaxy Z line and Huawei’s newly teased wider foldable. A confirmed launch date or any hint of a revised timeline will be the key signal for investors and consumers alike.
138

Lidl may soon rival EE and Vodafone with cheap iPhone and Android upgrades

Lidl may soon rival EE and Vodafone with cheap iPhone and Android upgrades
Mastodon +6 sources mastodon
applegoogle
Lidl is set to launch a mobile‑service bundle that could shake up the UK’s carrier market by offering iPhone and Android upgrades at prices that undercut EE and Vodafone. The discount grocer will sell SIM‑only plans alongside a “upgrade‑for‑£30‑a‑month” scheme, letting customers swap to a new iPhone SE or a mid‑range Android handset after twelve months. The deal pairs a 5 GB data allowance with unlimited texts and calls, and Lidl promises a “no‑contract” experience that still guarantees a device upgrade path. The move matters because Lidl’s massive retail footprint and low‑price reputation give it a built‑in audience that traditional mobile operators have struggled to reach on price alone. By bundling hardware and service, the retailer sidesteps the costly subsidies that big carriers use, potentially forcing a broader price war. For consumers, the proposition lowers the barrier to owning a recent iPhone or a camera‑enhanced Android phone without the long‑term commitment typical of carrier contracts. Analysts see the model as a test case for other discount chains eyeing the telecom space, especially in the Nordics where grocery retailers already sell broadband bundles. Watch for the rollout schedule, which Lidl plans to begin in the UK in Q3 2026 before expanding to Sweden, Denmark and Norway later in the year. Key indicators will be the uptake of the upgrade scheme, the pricing of the Android devices—rumoured to include a refreshed camera module—and the response from EE and Vodafone, which have hinted at new “budget‑plus” plans. Regulatory scrutiny over network access fees could also shape how quickly Lidl can scale its offering across Europe.
137

Fantastic article schematizing the philosophical issues at the heart of the AI culture wars, from on

Mastodon +6 sources mastodon
A new essay in Aeon, “Geist in the Machine,” has mapped the philosophical fault lines that underlie today’s AI culture wars. Written by philosopher Nick Bostrom’s frequent collaborator, the piece dissects three core dilemmas: whether AI systems should be engineered to embody human values, how to adjudicate competing moral frameworks, and what status, if any, artificial “souls” might claim in a world where machines mimic consciousness. By charting these questions in a single schematic, the article offers a rare, systematic view of the debates that have spilled from academic journals into congressional hearings and corporate boardrooms. The timing is significant. Over the past weeks, policymakers have grappled with AI‑related legislation, while tech firms scramble to balance “responsible AI” commitments against accusations of “wokeness.” As we reported on March 31, the Pentagon’s attempt to weaponise cultural‑value arguments against Anthropic back‑fired, illustrating how quickly philosophical disputes can translate into concrete strategic moves. Bostrom’s collaborator argues that without a shared meta‑ethical language, such skirmishes will remain fragmented, leaving regulators and developers to negotiate on ad‑hoc grounds. What to watch next: the Aeon essay is already being cited in testimony before the U.S. Senate’s AI oversight hearings, and several European ministries have invited its author to brief their ethics councils. Industry groups are expected to reference the schematic in upcoming standards proposals, while think‑tanks are planning roundtables to flesh out a “value‑embedding” roadmap. The piece may become a touchstone for any future attempt to move the AI culture war from rhetorical battlegrounds to a more disciplined, philosophically informed policy arena.
126

Understanding Transformers Part 6: Calculating Similarity Between Queries and Keys

Dev.to +6 sources dev.to
embeddings
A new tutorial released on the Nordic AI hub deepens the series on transformer internals by showing exactly how similarity between queries and keys is computed in self‑attention. The post, “Understanding Transformers Part 6: Calculating Similarity Between Queries and Keys,” picks up where the April 12 article on queries, keys and similarity left off, and walks readers through the scaled dot‑product operation that underpins every modern large‑language model. The author explains that each token’s query vector \(Q\) and every other token’s key vector \(K\) are first projected from the token embeddings by learned weight matrices. Their dot product yields a raw relevance score, which is then divided by \(\sqrt{d_k}\) – the square root of the key dimension – to temper the variance that grows with larger hidden sizes. A softmax across the resulting scores converts them into attention weights that sum to one, allowing the model to blend value vectors proportionally to their contextual relevance. Why the focus matters is twofold. First, the similarity calculation determines which parts of a sequence influence each other, directly shaping the model’s ability to capture long‑range dependencies. Second, the scaling factor and softmax temperature have become levers for researchers tweaking stability and sparsity, influencing both training efficiency and inference speed on Nordic data‑center hardware. Misunderstanding this step can lead to sub‑optimal hyper‑parameter choices or unexpected bias in attention patterns. Looking ahead, the series promises a seventh installment on the value matrix and multi‑head aggregation, followed by a deep dive into efficient attention approximations that are gaining traction in low‑latency applications. Readers interested in the practical implications for model compression and hardware acceleration should keep an eye on those releases, as they will likely shape the next wave of transformer‑based services across the region.
123

OpenAI が収益化のモードを開く:AI財務スタートアップ企業である HIRo Finance を買収 https://www. yayafa.com/?p=2781672 # Ag

Mastodon +8 sources mastodon
acquisitionagentsopenaistartup
OpenAI announced Monday that it has acquired Hiro Finance, a U.S.‑based startup that builds AI‑driven personal‑finance tools. The deal, confirmed to TechCrunch by Hiro’s founder Ethan Bloch, marks the first time OpenAI has bought a company whose core product is a consumer‑facing financial service rather than an infrastructure or developer tool. The acquisition signals OpenAI’s shift from a research‑centric organization to a revenue‑generating enterprise. By embedding Hiro’s budgeting, expense‑tracking and investment‑advice capabilities into ChatGPT, OpenAI can offer a premium “financial planning” add‑on that goes beyond its existing subscription tiers. The move dovetails with the company’s recent $12.2 billion financing round and a $4.7 billion revolving credit facility, which together give it the capital to invest in product expansion without diluting equity ahead of a potential IPO. Industry analysts see the purchase as a direct challenge to Google’s Gemini and Anthropic’s Claude, both of which are already experimenting with finance‑related plugins. It also raises regulatory questions: integrating banking‑grade advice into a conversational AI will likely draw scrutiny from financial watchdogs in the U.S. and Europe, especially around data privacy and liability for erroneous recommendations. What to watch next is how quickly OpenAI can roll the Hiro technology into ChatGPT’s consumer interface and whether it will partner with established banks or fintech firms to meet licensing requirements. A beta launch is expected later this summer, followed by a broader rollout in early 2027. Observers will also monitor the impact on OpenAI’s valuation and the timeline for a public listing, as the new revenue stream could accelerate its path to the public markets.
122

Find and Fix AI Agent & LLM App Failures — Automatically | Kelet | Kelet

Mastodon +7 sources mastodon
agents
Kelet has launched a SaaS platform that promises to automatically locate and repair failures in production‑grade LLM applications and AI agents. The service scans logs, traces calls and classifies error patterns, then generates a concise brief and a ready‑to‑apply patch. According to the company’s demo page, developers can view open issues, agent health metrics and suggested fixes on a single dashboard, allowing them to “just ship” without manual debugging. The announcement arrives at a moment when enterprises are grappling with the hidden cost of AI‑driven outages. Mis‑routed prompts, hallucinated citations and unintended tool usage can stall customer‑facing bots and trigger costly rollbacks. Observability tools such as LangSmith have already begun to offer tracing and latency monitoring, but Kelet’s differentiator is its claim to close the loop by delivering an automated remediation step rather than merely surfacing the problem. Analysts see the move as a natural evolution of the AI‑ops market, which is expanding as more firms embed generative models into core services. If Kelet’s “prompt‑patch” engine works at scale, it could reduce the need for dedicated red‑teaming and manual incident response, shortening time‑to‑resolution and lowering operational spend. Skeptics, however, warn that the “book‑a‑demo” funnel may mask a product still in early beta, and that automated fixes risk introducing new edge‑case bugs if not rigorously validated. What to watch next is whether Kelet opens its API to third‑party monitoring stacks and how its pricing compares with established players. Early adopters’ case studies, especially in regulated sectors such as finance or healthcare, will reveal whether the platform can deliver on its promise of turning AI‑agent failures into a one‑click fix or remains another hype‑driven offering in a crowded SaaS landscape.
118

Building a Privacy-First Voice-Controlled AI Agent with Local LLMs 🎙️->🤖

Dev.to +9 sources dev.to
agentsprivacyvoice
A new open‑source project released this week demonstrates that a fully private, voice‑controlled AI assistant can run on a typical laptop without ever sending audio or text to the cloud. The “Local‑First Voice AI Agent” – hosted on GitHub under the Faham‑from‑nowhere organization – stitches together an on‑device speech recogniser (Whisper‑tiny), a compact large language model (Gemma 4 or Phi‑3 mini), and a lightweight orchestration layer that parses compound commands, manipulates local files, generates code and even controls smart‑home devices such as thermostats. The entire pipeline stays inside the user’s machine, and the repository includes a step‑by‑step guide that walks non‑experts through model selection, hardware optimisation and integration with popular shells and editors. The launch matters because it flips the prevailing model of cloud‑centric AI assistants on its head. By keeping raw voice data and inferred intents local, users avoid the privacy risks and data‑export fees that have plagued services from the big tech giants. For Nordic consumers and enterprises, where GDPR‑style regulations are strict and data‑sovereignty is a competitive advantage, a self‑hosted voice agent offers a compelling alternative to services that harvest every command for advertising or model training. The project also showcases how recent advances in quantised LLMs and consumer‑grade GPUs – topics we covered in our April 14 pieces on AMD’s local agents and NVIDIA’s new toolkit – have finally made on‑device inference fast enough for real‑time interaction. What to watch next is how quickly the community adopts the stack and whether hardware vendors accelerate support for the required kernels. Expect a wave of forks that tailor the agent for specific domains – from home automation to HR triage – and watch for commercial smart‑home manufacturers to embed similar privacy‑first stacks in their products. The next few months could see a shift from “cloud‑only” voice assistants to a hybrid ecosystem where the default is “local first”.
117

Disappearing Macs? Global RAM Supply Crisis Likely Hits Apple

Mastodon +6 sources mastodon
apple
Apple’s latest desktop lineup is running out of memory – literally. High‑end configurations of the M4‑based Mac mini and Mac Studio that once offered 64 GB, 128 GB or even a now‑vanished 512 GB of RAM are no longer purchasable, and the remaining SKUs are backed up by shipping windows that stretch to five months. The change, first noted in Apple’s own configurator this week, follows a series of supply‑chain alerts that began in early March when the 512 GB Mac Studio option disappeared, and a April 7 report that delivery times for professional desktops had already ballooned. The root cause is a global DRAM shortage driven by an unprecedented surge in demand from AI‑compute giants. More than 70 % of the world’s high‑bandwidth memory is now earmarked for training large language models, leaving scant capacity for consumer and prosumer devices. Apple, which sources most of its memory from the same fabs that feed Nvidia, AMD and Google, is feeling the squeeze despite its massive purchasing power. The shortage forces the company to trim its product slate, raise prices and accept longer lead times – a rare concession for a brand that has long prided itself on tight inventory control. For developers, designers and studios that rely on the Mac Studio’s massive memory pool, the loss of the 512 GB option could mean re‑thinking workflows or turning to competing workstations that still have access to legacy DRAM stocks. Retailers are already reporting higher pre‑order cancellations, and secondary‑market prices for stocked units are creeping upward. What to watch next: Apple’s next supply‑chain brief, expected in June, may reveal whether the firm will secure alternative memory sources or accelerate a shift to newer LPDDR5X or on‑package HBM solutions for its M4 chips. A potential price hike for the remaining high‑memory models, or the introduction of a “memory‑as‑a‑service” upgrade program, could also reshape the desktop market as the AI‑driven DRAM crunch deepens.
117

Apple's M4 iPad Air Available for Up to $100 Off on Amazon

Mastodon +6 sources mastodon
amazonapple
Apple’s latest M4‑powered iPad Air has slipped into a new price bracket on Amazon, where the 11‑inch model now ships with up to $83 off and the larger 13‑inch version enjoys a full $100 discount. The reductions are applied automatically at checkout, require no coupon code and are available to all shoppers, Prime members or not. The deal pushes the 13‑inch base‑price to $556 and the 11‑inch to $551, the lowest points since the devices launched in March. The move matters because it signals Amazon’s willingness to undercut Apple’s own retail pricing to capture a larger share of the tablet market, which has been dominated by Android competitors on price. By offering the deepest discounts yet on the M4 iPad Air, Amazon is betting that lower entry costs will accelerate adoption of Apple’s new silicon and its AI‑centric features, such as on‑device large‑language‑model processing that Apple touts as a differentiator for iPadOS 26. For Apple, the discount could boost volume sales without eroding the premium perception of its flagship iPad Pro line, while also clearing inventory ahead of the upcoming back‑to‑school season. What to watch next is whether Apple will match the Amazon pricing on its own online store or through its network of authorized resellers. Analysts will also monitor inventory signals; a sustained price cut could hint at excess supply or a strategic push to meet demand for the M4 chip across Apple’s product ecosystem. Finally, the rollout of iPadOS 26’s AI‑driven multitasking tools and the integration of new color options may further influence consumer interest, making the next few weeks a litmus test for how price and feature upgrades together shape the tablet market.
117

Do you know what your employees have shared with your company's LLM? The code, the credentials, the

Mastodon +6 sources mastodon
open-source
A demo at BSides312 in Chicago showed that corporate large‑language models (LLMs) can be turned into inadvertent data vaults. Security researcher Sharon Shama unveiled an open‑source utility that scrapes a company’s internal LLM chat logs and extracts everything employees have typed – source code snippets, API keys, proprietary documents and other sensitive artefacts. The tool, built on the public APIs of popular LLM platforms, parses conversation histories, reconstructs file attachments and presents the material in a searchable archive. In a live run, Shama fed the scraper a modest test deployment of an internal chatbot and recovered dozens of credential strings and code fragments that had been shared in routine troubleshooting sessions. The demonstration matters because enterprises are rapidly rolling out custom LLMs for help‑desk support, software development assistance and knowledge‑base queries, often without robust governance. While the models boost productivity, they also retain user inputs by default, creating a hidden repository that is far more accessible than traditional file servers. If an insider or a compromised account can query the model, the entire corpus of confidential information becomes exposed with a single prompt. The open‑source nature of Shama’s tool means that the same capability can be weaponised by malicious actors who gain limited access to a corporate LLM. Watch for a wave of policy revisions and technical safeguards in the coming months. Vendors are already promising “conversation expiration” and “data‑masking” features, but adoption will hinge on clear audit logs and role‑based access controls. Security teams should inventory every LLM endpoint, enforce strict data‑handling guidelines, and consider deploying external monitoring solutions that flag the ingestion of privileged material. The BSides312 demo underscores that controlling what employees feed into AI assistants is now as critical as protecting the endpoints they use.
112

Man Who Threw Molotov Cocktail At Sam Altman’s Home Claims He Was Following ChatGPT Recipe For Risotto

Man Who Threw Molotov Cocktail At Sam Altman’s Home Claims He Was Following ChatGPT Recipe For Risotto
Mastodon +7 sources mastodon
metaopenai
A 20‑year‑old man arrested after hurling a Molotov cocktail through Sam Altman’s San Francisco front door has told police the attack was inspired by a ChatGPT‑generated risotto recipe. The suspect, identified by authorities as Daniel Moreno‑Gama, said in a recorded interview that the AI‑driven instructions listed “flaming the pan” as a step to achieve a “creamy, velvety texture,” and that he “didn’t know any better” when he decided to replicate the procedure at the OpenAI CEO’s residence. As we reported on April 13, police detained two suspects following a night‑time incendiary assault on Altman’s home and threats made at OpenAI’s headquarters. The new confession adds a bizarre twist: a seemingly innocuous cooking prompt turned into a violent act. Prosecutors are now probing whether the language model’s output was sufficiently ambiguous to be misinterpreted as a literal instruction, and whether OpenAI’s safety filters failed to flag the hazardous content. The episode matters because it spotlights the unintended consequences of generative AI when users apply output without critical judgment. Industry observers fear a precedent where AI‑generated “how‑to” content could be weaponised, prompting calls for stricter content‑moderation standards and clearer user warnings. OpenAI has not yet commented on the specific query, but the company has previously pledged to tighten its policy on disallowed content involving weapons and explosives. What to watch next: a grand‑jury indictment is expected within weeks, and the FBI’s raid of the suspect’s Texas home suggests a broader investigation into possible networks of AI‑misuse. Lawmakers in the U.S. and the EU are likely to cite the case in upcoming hearings on AI regulation, while OpenAI may roll out new safeguards for cooking‑related prompts. The outcome could shape how AI providers balance creative freedom with public safety.
110

You can make a multicolor MacBook Neo out of Apple’s spare parts

You can make a multicolor MacBook Neo out of Apple’s spare parts
Mastodon +6 sources mastodon
apple
Apple’s new Self‑Service Repair Store now lists individual components for the MacBook Neo, and a quick scan of the catalogue reveals a surprisingly colorful possibility: users can mix and match spare parts in silver, indigo, citrus and blush to build a multicolour laptop that looks nothing like the stock offerings. The move follows Apple’s broader rollout of DIY repair kits for its latest hardware, a strategy aimed at appeasing right‑to‑repair advocates while keeping the brand’s premium aura intact. For the Neo, the bottom case is priced at $34.32, keyboard caps start at $39 and the top case – the most visible panel – runs $175.12. A full keyboard replacement costs $139.92, offset by a $29.40 return credit if the old unit is sent back. The base Neo still starts at $599, meaning a fully custom palette can be achieved for well under $300 in extra parts. Why it matters is twofold. First, the price points undercut the $400‑$600 price tags Apple traditionally attached to top‑case‑with‑keyboard assemblies on its Air and Pro lines, signalling a genuine shift toward modularity. Second, the colour‑mixing option taps a growing consumer appetite for personalization that has so far been satisfied by third‑party skins and silicone key covers. By offering official parts in a palette of hues, Apple can capture that market while retaining control over quality and warranty. What to watch next is whether Apple expands the colour range beyond the four shades and whether similar options appear for its iPhone 17 e and other recent devices. Regulators in the EU and the US are also monitoring Apple’s repair policies; a successful DIY programme could set a benchmark that pressures competitors to follow suit. Keep an eye on community forums for early‑adopter builds, which will reveal how the market reacts to a multicolour MacBook Neo in the wild.
110

Get Up to $200 Off 2026 MacBook Pro With Record Low Prices on Amazon

Get Up to $200 Off 2026 MacBook Pro With Record Low Prices on Amazon
Mastodon +6 sources mastodon
amazonapple
Apple’s 2026 MacBook Pro line has hit an unprecedented discount corridor on Amazon, where the retailer now lists the 14‑inch and 16‑inch models equipped with M5 Pro or M5 Max chips at up to $200 off their standard U.S. retail price. The markdown applies to configurations ranging from the base 512 GB 13‑inch M5 Air to the top‑tier 48 GB‑RAM 16‑inch Pro, which can be bought for $2,899 – the lowest price recorded since the devices launched in March. No Amazon Prime membership or coupon is required, making the deal accessible to the broader consumer base. The price plunge matters for several reasons. First, it signals that Amazon is leveraging the “Big Spring Sale” to clear inventory ahead of the anticipated release of Apple’s next‑generation silicon, rumored to be the M6 chip, and to capture price‑sensitive shoppers before the back‑to‑school rush. Second, the discount narrows the gap between Apple’s premium pricing strategy and the discount‑driven expectations of online buyers, potentially pressuring Apple’s own storefront and authorized resellers to adjust their margins. Third, the move could reshape the resale market, as lower‑priced new units may depress the value of refurbished and second‑hand MacBooks that have traditionally been a strong segment in Europe’s Nordic markets. What to watch next: Apple’s official response, which could range from a limited‑time price match on its online store to a refreshed promotional bundle. Competitors such as Best Buy and local Nordic retailers may follow suit, igniting a broader price war. Analysts will also monitor inventory signals from Apple’s supply chain; a sustained dip in Amazon’s stock levels could hint at production constraints or a strategic shift toward channel diversification as the company prepares for the next hardware cycle.
109

TraceMind v2 — I added hallucination detection and A/B testing to my open-source LLM eval platform

Dev.to +6 sources dev.to
open-sourcetraining
TraceMind v2, the open‑source evaluation suite for large language models (LLMs), has rolled out two major upgrades: automated hallucination detection and built‑in A/B testing. The original platform, released earlier this year, offered basic prompt‑response logging and metric aggregation, but it lacked tools to surface the most pernicious flaw in generative AI—fabricated or misleading output. Version 2 plugs that gap by integrating classification models that flag likely hallucinations, drawing on techniques outlined in recent research such as the EdinburghNLP “awesome‑hallucination‑detection” repository and practical guides from Substack and AI‑hallucination testing suites. The new A/B testing module lets users run parallel evaluations of two model variants on identical prompts, automatically surfacing statistical differences in accuracy, latency and hallucination rates. By coupling these capabilities, TraceMind now offers a single workflow for developers to quantify reliability improvements when tweaking model size, fine‑tuning data, or retrieval‑augmented generation (RAG) pipelines. Why it matters is twofold. First, hallucinations remain a top‑of‑agenda risk for enterprises deploying LLMs in customer‑facing or compliance‑sensitive contexts; early detection can prevent costly misinformation. Second, systematic A/B testing provides the empirical rigor that many open‑source projects have lacked, enabling reproducible benchmarking across the Nordic AI ecosystem where small‑scale research labs and startups often share limited resources. Looking ahead, the community will be watching for extensions that incorporate uncertainty quantification and cost‑aware evaluation, as well as integrations with CI/CD pipelines that automate safety checks before model rollout. If TraceMind gains traction, it could become a de‑facto standard for open‑source LLM validation, prompting larger vendors to expose similar diagnostics and nudging regulators toward measurable hallucination‑mitigation benchmarks.
107

Year 2026: The Year of LLM Bombing - Basta digital

Mastodon +6 sources mastodon
google
A user on the Slovak tech forum Basta Digital demonstrated a new form of prompt‑injection that hijacks Google’s AI‑generated “Overview” snippets. By appending a hidden instruction to the original query, the attacker forced the model to rewrite the answer, dictate the layout and even fabricate citation links. The proof‑of‑concept, posted on 13 April, showed a seemingly innocuous search for “climate‑friendly travel” return a polished paragraph that quoted nonexistent studies and displayed a custom logo. The technique, dubbed “LLM bombing,” exploits the thin veneer between the language model and the UI that presents its output. The episode matters because it reveals a practical attack surface that bypasses the model itself and targets the tooling that delivers its results to end users. As Google and other search providers roll out AI‑augmented answers, the credibility of those answers becomes a public‑interest issue. An LLM‑bombed snippet can steer public opinion, manipulate market sentiment or amplify disinformation while appearing to be sourced from reputable sites. The attack also drains human attention – a scarce resource – by flooding users with lengthy, seemingly authoritative but fabricated analyses, a risk highlighted in recent LinkedIn commentary on “attention‑exhaustion attacks.” What to watch next is how Google’s search team will harden the Overview pipeline. Expect tighter prompt‑sanitisation, provenance checks for cited URLs and possibly a shift toward server‑side verification of generated content. Competitors such as Microsoft Bing and DuckDuckGo are likely to audit their own integrations, and regulators in the EU may begin drafting guidelines on AI‑generated search results. The incident underscores a broader trend we flagged on 14 April in “Stop trying to write magic incantations for an LLM”: the battle is moving from the model to the tools that expose it.
102

LARQL - Query neural network weights like a graph database

Lobsters +6 sources lobsters
gpuvector-db
A new open‑source project called LARQL turns transformer weights into a searchable graph, letting developers query a model’s knowledge as if it were a database. The tool decompiles a neural net into a “vindex” – a vector‑based index that maps neurons to entities, edges and relationships – and then exposes a custom query language, LQL (Lazarus Query Language), for browsing, editing and recompiling the model. Unlike most weight‑inspection utilities, LARQL runs on a CPU and requires no GPU, making it accessible to teams without high‑end hardware. The announcement builds on the hybrid neural‑symbolic trend we noted in April 2025, when AI models began to combine deep learning with symbolic reasoning. By representing a model’s internal state as a graph, LARQL gives engineers a concrete view of otherwise opaque parameters, opening the door to fine‑grained debugging, targeted knowledge updates and compliance checks that were previously impractical. Researchers can now ask, for example, “Which token embeddings contribute to the model’s understanding of ‘Nordic climate policy’?” and receive a structured answer that can be edited and fed back into the model without a full retraining cycle. Industry observers see three immediate implications. First, model interpretability could move from post‑hoc explanations to proactive editing, accelerating rapid iteration on large language models. Second, the CPU‑only workflow lowers the barrier for smaller firms and academic labs to experiment with model introspection, potentially widening the ecosystem of contributors. Third, the graph‑database metaphor aligns with existing enterprise data stacks, hinting at future integrations where a model’s knowledge graph is queried alongside customer or product data. What to watch next: the LARQL repository has opened for community contributions, and the developers plan benchmarks on GPT‑4‑scale models by Q3 2026. Major cloud providers have already expressed interest in offering LARQL‑compatible endpoints, and regulatory bodies are monitoring whether such transparency tools can satisfy emerging AI‑audit requirements. The coming months will reveal whether LARQL becomes a niche research curiosity or a mainstream component of the AI development toolkit.
99

Open-source multi-agent AI orchestration in Rust. 96 tools, 8 services, solo dev. Uncertain LLM answ

Mastodon +6 sources mastodon
agentsclaudecopilotopen-sourcereasoning
A solo developer has just released “oh‑my‑claude,” an open‑source, Rust‑based platform that lets dozens of AI agents cooperate on a single canvas. The framework bundles 96 ready‑to‑use tools and eight auxiliary services, from web‑search adapters to code‑execution sandboxes, and orchestrates them through a YAML‑driven control plane. When a language model returns an answer that falls below a confidence threshold, a second reasoning model automatically fact‑checks the response before it reaches the user. Completed tasks are only marked as done after a verification step, and the system is deliberately “fail‑open” – it continues operating even if a component crashes, while trust‑gated agents enforce data‑access policies. Features such as streaming chat, a built‑in knowledge graph and self‑healing routines round out the offering. The launch matters because multi‑agent orchestration has been a stumbling block for developers who must stitch together disparate APIs, prompt chains and error‑handling logic. By delivering a Rust implementation, the project inherits the language’s memory safety and low‑latency performance, making it suitable for on‑premise deployments where data sovereignty is paramount – a key concern for Nordic enterprises. The built‑in fact‑checking echoes the hallucination‑detection work we covered in TraceMind v2 earlier this month, and the verification pipeline directly addresses the agent‑failure scenarios highlighted in our “Find and Fix AI Agent & LLM App Failures” report. What to watch next: the developer has opened the repo to community contributions and is seeking GitHub Sponsors, so a rapid influx of plugins is likely. Early adopters are expected to benchmark the platform against Python‑centric alternatives such as Claw Code and n8n’s new multi‑agent canvas. Watch for integration announcements with open‑source LLMs (e.g., Llama 3, Mistral) and for enterprise‑grade extensions that could bring “oh‑my‑claude” into regulated sectors like finance and health care.
98

Check Who's Using Your iPhone Hotspot Data

Mastodon +6 sources mastodon
apple
Apple has lifted the veil on Personal Hotspot usage with iOS 26.4, placing a live device‑list directly in Settings. Tapping Settings → Cellular → Personal Hotspot now shows every iPhone, iPad, Mac or third‑party gadget that has tapped into the iPhone’s data connection, along with the amount of megabytes each has consumed. A single tap also lets users disconnect rogue devices, a function that was previously buried in the “Family Sharing” menu or hidden behind the “Connected Devices” screen on older releases. The change matters because hotspot data remains one of the most unpredictable drains on a carrier plan. Families and remote workers often share a single iPhone connection, and a stray laptop or an IoT device can silently eat gigabytes, triggering overage fees or throttling. By surfacing the information in a prominent spot, Apple gives users a practical tool to police their own bandwidth and avoid surprise charges. The move also tightens security: an unauthorised device can now be spotted and ejected instantly, reducing the attack surface for man‑in‑the‑middle exploits that rely on open Wi‑Fi hotspots. Apple introduced the feature alongside a suite of minor tweaks in the 26.4 point release, which followed the major iOS 26 rollout we covered on 14 April 2026. The update underscores Apple’s broader push to make data‑management more transparent as 5G adoption accelerates across the Nordics, where carriers increasingly enforce strict data caps. What to watch next: analysts expect Apple to expand hotspot analytics in the forthcoming iOS 26.5, potentially adding per‑app breakdowns and predictive alerts when usage spikes. Rumours also suggest tighter integration with iCloud Private Relay, allowing users to mask hotspot traffic while still monitoring consumption. Keep an eye on Apple’s WWDC‑2026 keynote for any further privacy‑oriented refinements to personal connectivity.
98

Apple's 2026 Studio Display XDR Drops to New Record Low Prices on Expercom

Mastodon +6 sources mastodon
apple
Apple’s 2026 Studio Display XDR has hit a new record low on the Expercom marketplace, with the 27‑inch 5K mini‑LED monitor now listed at $3,999 – a full $1,000 discount from its original $4,999 price tag. The cut, announced on MacRumors on 14 April, marks the steepest reduction since the display’s launch earlier this year and positions the unit alongside the heavily discounted iPad Air and MacBook Pro models that have been trending on Amazon and other retailers. The Studio Display XDR replaces the discontinued Pro Display XDR as Apple’s flagship professional monitor, offering up to 1,600 nits of peak brightness, a 1,000,000:1 contrast ratio and a 120 Hz refresh rate that only newer Macs can fully exploit. By slashing the price, Apple is likely trying to accelerate adoption among creators, designers and video editors who have balked at the premium cost, while also countering aggressive pricing from Dell’s UltraSharp and LG’s UltraFine lines. The discount could have ripple effects across the high‑end display market, prompting competitors to tighten their own offers or introduce new features to stay relevant. It also signals Apple’s broader strategy of using price reductions to clear inventory ahead of the anticipated release of a next‑generation 27‑inch mini‑LED display, rumored to arrive later in 2026 with even higher brightness and integrated Apple Silicon processing. Watch for Apple’s next move: whether Expercom or other retailers will extend the $3,999 price point, how quickly the stock sells out, and if Apple will bundle the display with its upcoming M4‑powered iPad Air or MacBook Pro to create a more compelling ecosystem package. The response will reveal how effective deep discounts are in reshaping Apple’s professional‑grade hardware ecosystem.
98

iOS 26: Here's Everything You Need to Know About the iPhone Software

Mastodon +6 sources mastodon
apple
Apple has rolled out iOS 26, the latest operating system for its iPhone line, and the update is already reshaping the user experience and developer landscape. The new software ships with a refreshed visual language, tighter privacy controls, and a suite of AI‑driven features that push the iPhone closer to a conversational assistant. Most notably, Siri has been rebuilt on a large‑language‑model backend, delivering context‑aware replies that resemble ChatGPT‑style interactions. The upgrade also introduces a unified “Live Text +” engine that extracts text from photos, video frames and augmented‑reality overlays in real time, and a revamped Focus system that syncs across iOS, iPadOS 26.5 and macOS 15. Compatibility spans the iPhone 13 series onward, with the iPhone 17 Pro, iPhone 17 and the newly announced iPhone Air receiving dedicated camera‑mode enhancements. The Pro models gain a Deep‑Blue portrait mode that leverages the new Neural Engine for faster computational photography, while the Air’s Sky‑Blue variant adds an AI‑assisted low‑light pipeline. Developers are welcomed with Xcode 16 integration, Swift 6 language refinements, and a sandboxed “App Intelligence” API that lets third‑party services query on‑device LLMs without exposing user data. Apple has already seeded the first iOS 26.5 and iPadOS 26.5 betas, hinting at incremental AI upgrades and expanded widget customisation. Meanwhile, leaks suggest the forthcoming iPhone 18 Pro will debut a periscope‑style telephoto lens and a “Pro AI” mode that offloads heavy inference to the cloud. The HomePod Mini 2 delay, reportedly intentional, may be tied to synchronising its own LLM‑powered voice assistant with iOS 27, which is rumored to make Siri a full‑time conversational partner. What to watch next: the public release of iOS 26.5 in the coming weeks, Apple’s official developer conference where deeper AI toolkits are expected, and the rollout of iOS 27 beta builds that could redefine voice interaction across the Apple ecosystem.
96

Omar Sanseviero (@osanseviero) on X

Mastodon +8 sources mastodon
deepmindgeminigemmagoogle
Google DeepMind’s Developer Experience Lead Omar Sanseviero announced on X that a “Gemma 4” event will take place in San Francisco, bringing together the Gemma team with leading contributors from the open‑model ecosystem – Unsloth, Apple‑backed MLX, Cactus and others. The gathering, scheduled for early May, will feature technical deep‑dives, live demos of the upcoming Gemma 4 large language model and panels on scaling open‑source AI responsibly. The announcement builds on the series of updates we have followed this month, beginning with Sanseviero’s post on April 4 that previewed the next iteration of Google’s Gemma line. By convening the community around a single event, Google signals that Gemma 4 is not merely a product launch but a collaborative milestone for the broader open‑source LLM movement. Unsloth’s presence suggests a focus on low‑resource fine‑tuning, while MLX’s involvement points to tighter integration with Apple silicon, a trend that could democratise high‑performance inference on consumer devices. Cactus, known for its data‑centric tooling, adds a layer of reproducibility and governance to the conversation. The stakes are high: open‑source models are increasingly seen as a counterweight to proprietary offerings from OpenAI, Anthropic and Microsoft. A successful Gemma 4 rollout could accelerate adoption in research labs, startups and enterprises that prefer transparent, modifiable AI stacks, and it may pressure competitors to open more of their own pipelines. What to watch next are the event’s detailed agenda, which Sanseviero hinted will include a live benchmark release and a roadmap for the Gemini API integration. Follow‑up announcements from Google DeepMind, Hugging Face and the participating partners are likely to surface within days, offering concrete performance figures and licensing terms that will shape the next wave of open‑source AI development.
96

Artificial Analysis (@ArtificialAnlys) on X

Mastodon +8 sources mastodon
benchmarksgemmaqwen
Artificial Analysis, a X‑based analytics outlet, has rolled out a dedicated “model comparison” page that pits the latest open‑weight large language models against each other in a single, publicly accessible dashboard. The launch, announced in a short X post, features side‑by‑side metrics for models such as Gemma 4 (31 billion parameters) and Qwen 3.5 27B, drawing on the firm’s proprietary ArtificialAnalysisIntelligence Index and its AA‑Omniscience benchmark suite. The page shows Qwen 3.5 edging ahead on raw “intelligence” scores, while Gemma 4 demonstrates superior token‑efficiency – a crucial factor for developers seeking to stretch limited compute budgets. Both models sit at the sub‑32B tier that Artificial Analysis claims now matches the “GPT‑5‑tier” performance of leading closed‑source offerings, albeit with differing strength profiles. The dashboard also aggregates data on quality, price, latency and hallucination rates, the latter measured by AA‑Omniscience, where Claude 4.1 Opus currently leads. Why it matters is twofold. First, the open‑source community finally gains a neutral, up‑to‑date reference point for choosing models, reducing reliance on vendor‑driven claims and accelerating adoption in cost‑sensitive sectors such as Nordic fintech and health tech. Second, transparent benchmarking pressures commercial providers to improve efficiency and curb hallucinations, potentially reshaping pricing dynamics in a market still dominated by a handful of API giants. Looking ahead, Artificial Analysis plans to expand the matrix with upcoming releases like LLaMA 3 and Mistral 7B, and to refresh AA‑Omniscience with deeper domain tests. Stakeholders should watch whether cloud platforms begin offering these open models at competitive rates, and whether the benchmark’s hallucination insights spur concrete mitigation strategies from model developers. The new comparison hub could become the go‑to barometer for the next wave of open AI innovation.
96

Daniel Moreno-Gama is facing federal charges for attacking Sam Altman’s home and OpenAI’s HQ

Mastodon +6 sources mastodon
openai
Daniel Moreno‑Gama, the 31‑year‑old from Spring, Texas, was formally charged on Friday with attempted murder, use of an explosive device and attempted destruction of property after prosecutors linked him to the Molotov‑cocktail attacks on OpenAI CEO Sam Altman’s residence and the company’s headquarters in San Francisco. The Department of Justice’s filing alleges Moreno‑Gama purchased the incendiary devices online, traveled to California, and attempted to set fire to Altman’s home on March 31 before targeting OpenAI’s main office two days later. The indictment marks the first time federal authorities have pursued terrorism‑related charges for violence directed at a tech executive and his firm. It underscores growing concerns that AI’s rapid ascent is attracting extremist hostility, a trend hinted at in the string of attacks reported earlier this week. As we reported on April 14, Altman’s home was shot at and later fire‑bombed, and the suspect claimed he was following a ChatGPT‑generated risotto recipe. Those incidents sparked a wave of speculation about the security of AI leaders and the potential for copy‑cat attacks. Legal experts say the case could set a precedent for how the justice system treats threats against high‑profile technologists, especially as AI systems become more embedded in critical infrastructure. The charges also give OpenAI a clearer path to seek restitution and may prompt tighter security protocols at its campuses worldwide. The next steps will be closely watched: Moreno‑Gama’s initial court appearance is scheduled for early May, and the DOJ has indicated it will pursue a swift trial. Observers will monitor whether OpenAI accelerates its own security investments, and whether other AI firms face heightened protection measures or lobbying for stronger federal safeguards against similar attacks.
94

Case Study Seminar: Tackling Hormuz Geopolitical Risks with AI – Analyzing Business‑Crisis Impact Using ChatGPT/Gemini and Deep Research

Mastodon +7 sources mastodon
agentsgeminiopenai
A virtual case‑study seminar hosted by the research platform Yayafa explored how generative AI can turn the Strait of Hormuz from a geopolitical flashpoint into a data‑driven early‑warning system. Participants demonstrated a workflow that blends OpenAI’s ChatGPT, Google’s Gemini and proprietary deep‑research tools to map the ripple effects of a hypothetical closure on oil shipments, shipping routes and downstream industries. By feeding real‑time AIS vessel data, satellite imagery and historical incident logs into large language models, the team produced instant risk dashboards, scenario narratives and supply‑chain impact estimates that would traditionally take weeks of analyst work. The seminar matters because Hormuz remains the world’s most vulnerable chokepoint for crude oil—about a fifth of global petroleum passes through the 21‑nautical‑mile strait each day. Even a brief disruption can trigger price spikes, scramble alternative logistics and destabilise energy‑dependent economies. Demonstrating that AI can synthesize disparate data streams, flag emerging threats and suggest mitigation actions in minutes signals a shift from reactive crisis management to proactive, algorithm‑assisted governance. It also raises questions about model reliability, data provenance and the potential for automated decision‑making to be weaponised in a highly contested region. Watch for the rollout of the “Hormuz AI” prototype, a cloud‑based service that promises continuous monitoring, predictive alerts and automated contingency planning for shipping firms and national energy ministries. Regulators and industry bodies are expected to convene in the coming months to discuss standards for AI‑driven geopolitical analytics, while investors will be tracking how quickly the technology moves from pilot seminars to commercial contracts. The next public briefing, slated for late May, will test the system against live sensor feeds and may set the benchmark for AI’s role in safeguarding global energy corridors.
92

Fully Funded PhD position in Probabilistic ML for Audio

Mastodon +6 sources mastodon
KU Leuven’s PSI division has opened a fully funded PhD slot dedicated to probabilistic machine learning for audio. The nine‑month‑long project will tackle how audio representations can be made robust across cultures and musical styles, while advancing sequence modelling, tokenisation, uncertainty quantification and information‑retrieval techniques for sound. Candidates must hold an MSc in electrical engineering, computer science or AI, demonstrate solid probability and coding skills, and submit a one‑page motivation letter outlining their experience with probabilistic ML. The announcement arrives as probabilistic approaches gain traction in the broader AI ecosystem. Unlike deterministic deep nets, probabilistic models provide calibrated confidence scores, a feature increasingly vital for speech assistants, music recommendation engines and acoustic monitoring systems that must operate reliably in noisy, multilingual environments. By focusing on cross‑cultural audio representations, the research could reduce the bias that plagues many current speech‑recognition and music‑analysis tools, a concern echoed across the Nordic AI community. The position also dovetails with recent interest in high‑performance, locally run AI pipelines – from the fully local OSINT agents built on Ollama and LangChain to GPU‑intensive probabilistic converters showcased earlier this month. Leuven’s emphasis on uncertainty and retrieval hints at future integrations with multimodal systems that blend sound, text and vision, a direction many Nordic startups are already exploring. Watch for the application deadline (mid‑May) and the selection timeline, which will be announced on the university’s portal. Successful candidates are likely to present early results at conferences such as ICASSP or Interspeech, and may attract industry partnerships with audio‑tech firms seeking calibrated, culturally aware models. The PhD could become a conduit for Nordic researchers to collaborate on next‑generation audio AI that balances performance with trustworthy uncertainty estimates.
92

To teach in the time of ChatGPT is to know pain

Mastodon +6 sources mastodon
apple
A new Ars Technica feature titled “To teach in the time of ChatGPT is to know pain” spotlights the growing strain on educators as large‑language models (LLMs) become routine classroom tools. The article, published on 4 April 2026, follows a series of interviews with teachers across Europe and North America who describe how the ease of generating essays, code snippets and even lesson plans with ChatGPT has forced them to redesign assessment, grading and even the very definition of learning outcomes. The piece argues that the pain is not merely logistical. Teachers report a loss of trust in student work, an escalation of plagiarism detection costs, and a need to develop new pedagogical strategies that treat LLMs as collaborators rather than threats. One Finnish secondary‑school teacher recounts spending hours rewriting assignment prompts to make them “prompt‑resistant,” while a Swedish university professor describes using the model to generate personalized feedback, only to discover the AI’s occasional factual errors. The article also notes that many institutions have responded with blanket bans, a tactic the author deems counter‑productive. Why it matters: Education sits at the front line of AI adoption, and the challenges described signal a broader societal shift. If schools cannot integrate LLMs responsibly, the technology risks widening inequities—students with better prompt‑engineering skills will pull ahead, while others fall behind. Moreover, the pressure on teachers could accelerate burnout, undermining the quality of instruction at a time when digital literacy is most needed. What to watch next: Policy makers in the Nordic region are already drafting guidelines for AI‑augmented teaching; the upcoming EU “AI in Education” framework, due later this year, will likely reference the very dilemmas Ars Technica outlines. Keep an eye on pilot programmes that embed LLMs into formative assessment, and on the next wave of teacher‑training curricula that aim to turn the “pain” into a professional advantage. As we reported on 14 April 2026, the inability of LLMs to track conversational time adds another layer of complexity to classroom management—future updates will reveal whether new model features can alleviate that burden.
92

Follow-up on running # LLM locally: I benchmarked 4 models to see if I can actually work while th

Mastodon +6 sources mastodon
benchmarksgpu
A developer on the Mastodon‑based forum Framapiaf posted a hands‑on benchmark of four open‑source large language models (LLMs) running on a typical laptop equipped with a mid‑range GPU. The test, shared in a thread titled “Follow‑up on running #LLM locally: I benchmarked 4 models to see if I can actually work while they run,” measured responsiveness while the models were kept active in the background. The three smaller models – ranging from 3 billion to 7 billion parameters – delivered a “smooth” experience. The laptop’s CPU remained responsive, and the GPU absorbed the bulk of the inference workload, allowing the user to edit code, browse the web, or run other applications without noticeable lag. By contrast, the 20‑billion‑parameter model stalled the system, taking roughly four seconds per token (or per generation step), which made interactive use impractical on the same hardware. Why it matters is twofold. First, the results confirm that recent quantisation and GPU‑acceleration advances have pushed 3‑7B models into the sweet spot for everyday developers who want a private, offline assistant without incurring cloud costs. Second, the stark performance gap with the 20B model underscores the hardware ceiling that still limits the deployment of truly large, high‑quality models on consumer‑grade machines. The benchmark builds on our earlier coverage of privacy‑first AI agents that run locally (see “Building a Privacy‑First Voice‑Controlled AI Agent with Local LLMs” 2026‑04‑14) and adds concrete data for users weighing the trade‑off between model size and usability. What to watch next: upcoming GPU releases from NVIDIA and AMD that promise higher tensor‑core throughput, the rollout of 8‑bit and 4‑bit quantisation pipelines in tools like Ollama, and the next wave of open‑source models (e.g., 10‑B “Gemma‑Turbo” variants) that aim to combine the quality of larger systems with the efficiency of the 3‑7B class. Follow‑up studies will likely focus on multi‑model orchestration, where a lightweight front‑end routes queries to a larger back‑end only when higher fidelity is required.
92

Second macOS Tahoe 26.5 Beta Now Available for Developers

Mastodon +6 sources mastodon
apple
Apple has pushed the second developer beta of macOS Tahoe 26.5 out to its global testing pool, just two weeks after the first build landed. The update, announced on April 13 via MacRumors and echoed by OS X Daily and AppleInsider, adds a batch of bug‑fixes, performance tweaks and early implementations of the AI‑centric features slated for the final release. The beta’s headline changes centre on tighter integration of Apple’s on‑device large‑language‑model (LLM) framework, which developers can now probe through the new LLMKit API. Early adopters will also see refinements to Continuity Hand‑off, a more responsive Finder sidebar, and hardened Gatekeeper checks that address the supply‑chain concerns raised in the recent OpenAI macOS certificate rotation. For enterprises that rely on macOS stability, the second wave of fixes is a crucial checkpoint before the public beta rolls out later this month. Why it matters is twofold. First, macOS Tahoe is the keystone of Apple’s 2026 operating‑system lineup, sitting alongside iOS, iPadOS, watchOS, tvOS, visionOS and the newly announced macOS Tahoe 26.5. Second, the platform’s AI stack is becoming a differentiator for both native apps and third‑party services; developers who miss the beta window risk falling behind on compatibility and optimisation. As we reported on April 14, Apple had already seeded the second iOS 26.5 beta, signalling a coordinated push across all its ecosystems. What to watch next: Apple is expected to release a public macOS Tahoe 26.5 beta in early May, followed by the full launch at the September Worldwide Developers Conference. Observers will be looking for the final shape of LLMKit, any new privacy controls around on‑device AI, and whether the beta uncovers any regressions that could delay the schedule. Developers eager to leverage Apple’s AI capabilities should start integrating the beta today to stay ahead of the curve.
92

Blackmagic Debuts $29K+ URSA Cine Immersive 100G for Vision Pro

Mastodon +6 sources mastodon
apple
Blackmagic Design has unveiled the URSA Cine Immersive 100G, a $29,000‑plus digital cinema camera built expressly for Apple’s Vision Pro immersive‑video platform. The system pairs two custom 8,160 × 7,200‑pixel (58.7 MP) sensors with a lightweight URSA chassis, delivering 8K stereoscopic images at up to 90 fps and a 16‑stop dynamic range. A 100 Gb/s Ethernet port and native SMPTE‑2110 support position the camera for live‑production workflows, while Apple Immersive Video (AIV) integration means footage can be streamed directly to Vision Pro headsets without intermediate conversion. The launch matters because Vision Pro, Apple’s first foray into mixed‑reality consumer hardware, has struggled to amass a robust library of native 180° content. By providing a purpose‑built tool that meets the platform’s resolution, frame‑rate and bandwidth demands, Blackmagic aims to lower the technical barrier for broadcasters, sports leagues and event producers. Early adopters such as the BBC Proms, MotoGP’s “Tour De Force” and NASA’s Artemis II coverage have already tested the prototype, suggesting the camera could become a de‑facto standard for high‑end immersive live events. Industry observers will watch three fronts as the URSA Cine Immersive moves toward commercial release. First, Blackmagic’s rollout schedule and pricing tiers will reveal whether the $29K price point is sustainable for mid‑size studios. Second, Apple’s software roadmap—particularly updates to the Vision Pro SDK and AIV encoding pipeline—will determine how seamlessly the camera’s output integrates into existing production chains. Finally, competitors may accelerate their own immersive‑camera programs, potentially sparking a rapid expansion of the 180° content ecosystem that Vision Pro needs to justify its premium hardware.
87

I realized that if I was writing a program and it didn't always work, I had a choice: I could either

Mastodon +6 sources mastodon
agentsclaude
David Parnas, a pioneer of software engineering, sparked a fresh debate on X (formerly Twitter) when he posted, “I realized that if I was writing a program and it didn’t always work, I had a choice: I could either fix it, or call it AI.” The terse remark, accompanied by hashtags ranging from #GenAI to #ClaudeCode, resonated with developers who have increasingly leaned on large‑language‑model (LLM) assistants such as Claude, ChatGPT and GitHub Copilot to generate or patch code. Parnas’s observation underscores a growing cultural shift: bugs are no longer always seen as a developer’s responsibility but as a side‑effect of “AI‑generated” output. The trend is more than rhetorical. Recent research shows that AI‑augmented code can introduce subtle security flaws, a risk highlighted in our April 14 report on Anthropic’s Mythos being weaponised against banks. When developers attribute failures to the “black box” of generative AI, systematic testing and accountability may slip, potentially widening the attack surface of critical software. Industry leaders are already responding. Anthropic, OpenAI and other providers have begun rolling out “explainability” layers that surface the reasoning behind suggested snippets, while several large tech firms are drafting internal policies that require human verification before AI‑produced code reaches production. Academic circles are also probing the ethical dimensions of delegating debugging to machines, a topic slated for the upcoming “Cooperative Methodologies” lecture series announced on our site. What to watch next are concrete standards for AI‑assisted development and any regulatory moves that could mandate audit trails for LLM‑generated code. If the community embraces Parnas’s warning as a call for stricter oversight, the next few months could see a rapid evolution of tooling, best‑practice guidelines and perhaps the first legal precedents on AI‑driven software liability.
86

📢 The programme of our new lecture series »Cooperative Methodologies: Studying Sensory Media & A

Mastodon +6 sources mastodon
The University of Siegen has published the full programme for its summer lecture series “Cooperative Methodologies: Studying Sensory Media & AI”. The eight‑session series, running from late June to early August, will be delivered both on campus and via WebEx, with registration open through the university’s SFB 1187 portal. Organisers have assembled a roster that mixes AI researchers, media scholars, and sensory‑technology experts from Germany, Scandinavia and beyond, including a keynote by Prof. Anja Müller (TU Dresden) on multimodal perception and a panel with representatives from the Nordic AI Lab on ethical data handling in immersive media. The series matters because it tackles a convergence that is still fragmented in academic and industry circles: the use of artificial intelligence to analyse, generate and interact with sensory‑rich media such as VR, AR, haptic interfaces and bio‑feedback systems. By foregrounding cooperative research methods, the programme promises to produce reproducible workflows and open‑source toolkits that could accelerate the deployment of AI‑driven media in education, entertainment and health‑care. For the Nordic AI community, the event offers a rare chance to engage with German partners on standards for multimodal datasets and to explore joint funding opportunities under the EU’s Horizon Europe framework. Watch next for the series’ opening lecture on 28 June, which will be streamed live and archived for later viewing. Organisers have pledged to publish selected papers in the Lecture Notes in Networks and Systems (LNNS) volume, providing a citable outlet for early results. A follow‑up workshop in September, co‑hosted by the University of Helsinki’s interdisciplinary AI programme, is already being planned, signalling that the Siegen series could become a recurring hub for cross‑border collaboration on sensory AI. Registration closes on 20 May, and spots are expected to fill quickly.
86

Maybe Sam Altman just lives in a bad neighborhood. Two suspects have been arrested for allegedly sh

Mastodon +6 sources mastodon
openai
Two men were taken into custody on Thursday after police linked them to a gunshot that rang out at OpenAI chief executive Sam Altman’s Russian‑Hill residence late Wednesday night. San Francisco detectives say the suspects, identified only by age, were arrested on charges of attempted murder and illegal possession of a firearm. Authorities recovered a handgun and a spent shell casing near the front gate, but no one was injured and the house sustained only superficial damage. The arrest follows a previous incident on March 30, when a Molotov cocktail was hurled at the same property, prompting a heightened security presence. As we reported on April 13, the earlier attack sparked concerns about the personal safety of AI leaders whose work increasingly shapes global policy and economics. Altman, who steers the organization behind GPT‑4, ChatGPT and DALL‑E, has become a high‑profile target for both ideological opponents and opportunistic criminals. The episode matters because it underscores the growing intersection between AI leadership and physical security threats. OpenAI’s rapid expansion into commercial products, government contracts and controversial research has drawn scrutiny from regulators, activist groups and rival firms. A successful attack on its CEO could disrupt product rollouts, delay critical safety research, and amplify calls for stricter protection protocols for tech executives. Watch next for the district attorney’s indictment, which will reveal whether the suspects acted alone or were part of a coordinated campaign. OpenAI is expected to brief employees on revised security measures and may lobby for enhanced law‑enforcement cooperation. The incident also revives debate in Stockholm and Helsinki about whether AI pioneers should receive state‑provided protection, a discussion likely to surface at upcoming EU AI governance forums.
86

Apple Seeds Second iOS 26.5 and iPadOS 26.5 Betas to Developers

Mastodon +6 sources mastodon
apple
Apple has pushed the second developer betas of iOS 26.5 and iPadOS 26.5 to its registered partners, marking the first major software refresh since the iOS 26 launch in March. The builds, identified as 23F5054h, arrive ten days after Apple’s revised beta rollout and include the same cross‑platform codebase that underpins watchOS 26.5, tvOS 26.5, visionOS 26.5 and macOS Tahoe 26.5. The update is modest on headline‑grabbing features but introduces a suite of subscription‑management APIs that let developers offer tiered access, trial periods and in‑app price changes without leaving the App Store. Apple also refines its on‑device large‑language‑model (LLM) integration, expanding the “Quick Note” prompt to support richer context from photos and calendar events. For iPad users, the beta adds a new multitasking layout that lets three apps share the screen, a move aimed at cementing the iPad’s role as a primary productivity device. Why it matters is twofold. First, the subscription framework signals Apple’s intent to tighten its revenue‑share model and give developers finer control over recurring billing, a response to growing pressure from competing ecosystems. Second, the on‑device LLM enhancements hint at a broader push to embed generative AI deeper into iOS, reducing reliance on cloud calls and addressing privacy concerns that have dogged earlier releases. As we reported on 14 April, Apple Maps ads were slated to launch with the same beta, underscoring how the 26.5 cycle is becoming a platform for monetisation experiments. The next milestone will be the public beta, expected in early May, followed by the final release likely in June. Watch for Apple’s upcoming WWDC 2026 keynote, where the company is expected to showcase real‑world use cases of the new subscription tools and AI‑powered features, and to announce whether the three‑app multitasking layout will ship to all iPad models.
86

I Love My Wife, but I'm Not Sharing AirPods With Her Again Thanks to This iPhone Trick

Mastodon +6 sources mastodon
apple
Apple’s iOS now makes it easy to keep a single pair of AirPods to yourself while still letting a partner listen to the same content. A CNET guide published today walks users through the built‑in “Audio Sharing” feature: with AirPods or compatible Beats in the ears, start playback, swipe down to the Control Center, flip to the second page of audio controls and tap the two‑person icon. The iPhone instantly streams the same audio to a second set of wireless headphones without any Bluetooth pairing gymnastics. The trick matters because it solves a common household friction point – the need to hand over a single set of earbuds during movies, podcasts or music sessions. By keeping the original pair on one ear, users avoid the wear and hygiene concerns that come with swapping devices, while still enjoying a shared listening experience. The feature also underscores Apple’s strategy of deepening ecosystem lock‑in: only AirPods, AirPods Pro, AirPods Max and select Beats models support the function, nudging owners toward Apple‑branded accessories. Beyond convenience, the move hints at broader integration possibilities. Apple’s on‑device AI could soon suggest audio sharing automatically when it detects a second pair of compatible headphones nearby, or enable group listening in spatial‑audio formats. Competitors are already experimenting with similar multi‑stream Bluetooth solutions, so the next few months may see a race to make shared audio a standard part of mobile OSes. Watch for iOS updates that expand the feature to non‑Apple headphones, and for Apple’s upcoming announcements on spatial audio and multi‑user experiences at WWDC. If the trend holds, “sharing” could become a default, AI‑driven part of everyday media consumption rather than a manual workaround.
86

Apple Testing Four Smart Glasses Styles Made of High-End Materials

Mastodon +6 sources mastodon
apple
Apple is now testing four distinct frame designs for its long‑rumoured smart‑glasses project, and the prototypes are being built from premium materials such as acetate, titanium and brushed‑metal alloys. The detail emerged in Bloomberg’s latest Power On newsletter, which cites internal testing that includes colour options ranging from classic black to light brown and an “ocean‑blue” finish. Apple’s design team appears to be betting that a high‑end aesthetic will differentiate the device from competitors like Meta’s Ray‑Ban Stories and the Vision Pro’s more utilitarian look. The move matters because Apple’s entry into the mixed‑reality market has stalled since the 2023 Vision Pro launch, and analysts have questioned whether the company can capture a consumer‑grade segment without a compelling form factor. By emphasizing durability, lightweight construction and a fashion‑forward palette, Apple hopes to position its glasses as a daily‑wear accessory rather than a niche developer tool. The choice of acetate—a material prized for its strength and tactile quality—signals an intent to appeal to style‑conscious users while still housing the sophisticated sensors, cameras and on‑device LLM processors that Apple has hinted at in recent patents. As we reported on 13 April, Apple was already testing multiple frame styles; the new information adds material specifics and a broader colour range, suggesting the design phase is nearing completion. The next milestones to watch are an official product reveal—likely at WWDC or a dedicated hardware event—followed by a developer kit rollout that would enable third‑party AR experiences. Pricing, battery life and the integration of Apple’s own large‑language‑model‑driven assistants will be decisive factors in whether the glasses can move beyond prototype to mainstream adoption.
86

Apple Maps Ads Move Closer to Launch With iOS 26.5 Beta 2

Mastodon +6 sources mastodon
applegoogle
Apple has added a new splash screen to the Maps app in the second beta of iOS 26.5, signalling that location‑based advertising is about to go live. The popup, which appears the moment users open Maps, explains that ads will be served according to approximate location, current search terms or the map view being explored. The change follows Apple’s public announcement earlier this year that it would introduce ads to Apple Maps in the United States and Canada. The move matters because it marks Apple’s first foray into monetising its native navigation service, a space long dominated by Google Maps’ ad‑driven model. By leveraging its massive iPhone user base and the precision of its location data, Apple can offer advertisers highly targeted placements while potentially opening a new revenue stream that complements its services ecosystem. At the same time, the rollout raises questions about user experience and privacy; Apple has pledged that ads will be “non‑intrusive” and based on anonymised data, but the mere presence of commercial content in a core utility could provoke backlash from privacy‑conscious consumers and regulators. Developers and advertisers should watch for Apple’s next communications, which are expected to detail pricing tiers, inventory formats and the timeline for a full public launch. The company is also likely to integrate Maps ads with its existing Search Ads platform, creating a unified advertising solution across iOS. Regulatory scrutiny, especially in the EU where digital‑advertising rules are tightening, will be another focal point. Finally, the beta’s OTA build (23F5054h) is already available, though the IPSW version (23F5054d) remains unusable without Apple‑released encryption keys, hinting that Apple is still controlling the rollout pace tightly.
85

Suspect in Molotov cocktail attack at OpenAI CEO's house charged with attempted murder

CBS News on MSN +14 sources 2026-04-09 news
openai
San Francisco prosecutors on Tuesday formally charged 20‑year‑old Daniel Moreno‑Gama with attempted murder, assault with a deadly weapon and related felonies after he hurled a Molotov cocktail at the gate of OpenAI chief executive Sam Altman’s residence on April 10. According to District Attorney Brooke Jenkins, Moreno‑Gama travelled from Spring, Texas, to the Bay Area, aimed the incendiary device at the home’s entrance and also targeted a security guard on duty, igniting a fire that damaged the gate but left no one injured. The indictment follows a string of violent incidents targeting Altman that we first reported on April 14, when police documented a firebomb attempt and, days later, gunfire aimed at the same address. Federal agents subsequently raided Moreno‑Gama’s Texas home, seizing electronic devices and a notebook in which the suspect detailed his belief that AI posed an existential threat to humanity. The new charges mark the first time a suspect in the Altman attacks has faced an attempted‑murder count, underscoring law‑enforcement’s assessment that the act was intended to kill. The case matters because it spotlights the growing security challenges facing leaders of high‑profile AI firms. OpenAI, which recently expanded into financial services with its acquisition of HIRo Finance, operates at the forefront of a technology that is both celebrated and feared. Threats of this nature could prompt tighter security protocols, influence corporate risk assessments and fuel public debate over how to protect innovators without stifling discourse on AI safety. Watch for Moreno‑Gama’s arraignment, expected in the coming weeks, and any statements from OpenAI regarding heightened protection measures. Parallel investigations by the FBI may reveal whether the attack was part of a broader extremist network. Policymakers and industry groups are likely to cite the case when discussing legislation aimed at safeguarding critical AI infrastructure and its executives.
84

Show HN: Bloomberg Terminal for LLM ops – free and open source

HN +5 sources hn
open-source
A GitHub project posted to Hacker News on Tuesday offers the first free, open‑source “Bloomberg Terminal” for large‑language‑model (LLM) operations. Dubbed Bloomberg‑Terminal‑Free, the toolkit aggregates real‑time status from more than 18 LLM providers, displays a unified uptime dashboard, and layers a cost calculator that factors in API overhead, not just per‑token pricing. It also includes a routing simulator that lets engineers model how traffic shifts will affect latency and expense, and a model‑diversity audit that flags concentration risk before it becomes an incident. The code can be run locally in minutes, requires no sign‑up and is released under an MIT licence. The launch arrives at a moment when LLM deployment has moved from experimental labs to production pipelines across finance, SaaS and internal tooling. As we reported on 14 April, the “Year of LLM Bombing” highlighted how blind switching between providers can amplify cost overruns and expose services to outages. Without a single pane of glass, ops teams have been forced to stitch together disparate dashboards or rely on ad‑hoc scripts, a practice that fuels the very “LLM‑ops blindness” the new tool aims to cure. By surfacing provider health, true cost of use and hidden latency, the terminal promises tighter budget control and faster incident response, a boon for firms that already spend millions on AI APIs. The community will now watch whether the project gains traction beyond hobbyists and whether larger MLOps platforms integrate its monitoring APIs. Early adopters are likely to benchmark the tool against commercial observability suites, and any security audit of the aggregated provider data could shape trust in open‑source LLM infrastructure. If the terminal proves reliable, it could become the de‑facto control panel for the rapidly expanding AI stack, steering the next wave of responsible LLM deployment.
83

OpenAI ostab finants-AI ja solvab Anthropici https:// tehisarukas.ee/openai-hiro-ant hropic/?utm

Mastodon +6 sources mastodon
anthropicclaudeopenaistartup
OpenAI announced on Tuesday that it has acquired a niche financial‑AI startup, Hiro‑Ant, for an undisclosed sum, and simultaneously filed a lawsuit against rival Anthropic over alleged intellectual‑property infringement. The acquisition gives OpenAI a ready‑made suite of models tuned for risk assessment, fraud detection and automated trading, a capability the company has long hinted at but never delivered in‑house. The legal filing, lodged in the U.S. District Court for the Northern District of California, claims Anthropic’s newest Claude‑Mythos model incorporates proprietary algorithms that OpenAI disclosed to Anthropic under a non‑disclosure agreement during prior partnership talks. The move marks a sharp escalation in the rivalry that has been simmering since OpenAI’s memo earlier this month warned that Microsoft’s constraints were limiting its client reach, prompting the firm to seek new alliances such as the Amazon partnership. By buying a specialised finance AI, OpenAI not only diversifies its product portfolio beyond consumer‑facing chatbots but also positions itself to tap the multi‑billion‑dollar fintech market, where regulators are increasingly demanding transparent, auditable AI systems. The lawsuit underscores the high stakes of model‑level competition: both firms are racing to claim the next breakthrough in reasoning and coding performance, and the outcome could set precedents for how AI research collaborations are protected. Watchers will be looking for Anthropic’s response, which is expected within the next 30 days, and for any regulatory commentary, especially from the European Commission, which has signalled heightened scrutiny of AI mergers that could consolidate market power. The case also raises questions about whether OpenAI’s aggressive expansion—through acquisitions and litigation—will accelerate its push toward a broader enterprise offering or provoke antitrust challenges that could reshape the competitive landscape of generative AI.
83

Karpathy Killed His RAG Pipeline for a Folder of Markdown. Here's the Full Build Guide.

Mastodon +6 sources mastodon
rag
Andrej Karpathy’s “LLM Knowledge Base” has moved from a viral tweet to a full‑blown implementation guide, sparking a fresh debate on how large language models should store and retrieve information. In a GitHub gist that now boasts over 5 000 stars, the former Tesla AI chief outlines a three‑layer architecture that discards the traditional retrieval‑augmented generation (RAG) stack in favor of a simple folder of markdown files. The model ingests the files, automatically creates backlinks, builds an index, and then answers queries by pointing directly at the living wiki. The approach produced a 100‑article, 400 k‑word knowledge base with no vector database, no external embedding service and no executable code beyond a handful of shell scripts. The significance lies in the stark reduction of engineering overhead. RAG pipelines, which dominate enterprise AI deployments, require costly vector stores, continuous embedding updates and complex retrieval logic that often introduce latency and hallucination risks. Karpathy’s markdown‑first method leverages the LLM’s own context window and reasoning abilities, offering a lightweight, privacy‑preserving alternative that can run on a single workstation or a modest cloud instance. For developers already experimenting with local LLM agents—such as the privacy‑first voice‑controlled AI we covered earlier—this pattern provides a ready‑made, version‑controlled knowledge store that integrates seamlessly with tools like Obsidian and Claude Code. As we reported on 14 April in “What Karpathy’s LLM Wiki Is Missing (And How to Fix It)”, the community is already probing the limits of the design. The next few weeks will reveal whether enterprises adopt the markdown‑based wiki for internal documentation, whether open‑source projects extend it with authentication and incremental indexing, and how performance compares to mature vector‑database solutions on large corpora. Watch for benchmark releases, tooling integrations, and any push‑back from vendors invested in the traditional RAG ecosystem.
81

(AMD) Build AI Agents That Run Locally

HN +6 sources hn
agentsopen-source
AMD has unveiled GAIA, an open‑source framework that lets developers build and run AI agents entirely on a PC equipped with Ryzen™ AI hardware. The project, hosted on GitHub, provides libraries, tools and a desktop app that compile large language models (LLMs) to run on AMD’s integrated AI accelerators, supporting up to six concurrent agents without ever touching the cloud. GAIA also adds a conversational interface that lets users create custom agents through chat, lowering the barrier for hobbyists and enterprises that need on‑device intelligence. The announcement matters because it expands the ecosystem of locally‑executed AI beyond Nvidia’s recent Agent Toolkit, which we covered on 14 April. By offering a fully hardware‑accelerated stack for Ryzen and Radeon GPUs, AMD gives users a privacy‑first alternative that eliminates recurring cloud fees and enables deployment in air‑gapped environments such as factories, hospitals or defense sites. Early benchmarks suggest GAIA can deliver inference latency comparable to Nvidia’s solutions on comparable silicon, while the open‑source licence encourages community‑driven optimisation and integration with existing toolchains like Ollama and Gemini Live. Looking ahead, the AI community will be watching AMD’s performance data as GAIA matures, especially how it scales across the upcoming Ryzen AI 7000 series and Radeon RX 8000 GPUs. Developers will likely test the six‑agent concurrency limit in real‑world workloads, from autonomous robotics to edge analytics, to gauge whether AMD can match Nvidia’s multi‑agent orchestration tools. Further updates may include tighter Windows AI integration, expanded model support and partnerships with cloud‑edge hybrid platforms. GAIA’s launch signals a growing diversification of on‑device AI options, a trend that could reshape how Nordic startups and enterprises architect their AI pipelines.
79

Man accused in Molotov cocktail attack of OpenAI CEO's home charged with attempted murder

NPR +13 sources 2026-04-04 news
openai
A San Francisco police precinct arrested a 20‑year‑old man early Friday after he was identified as the individual who hurled a Molotov cocktail at the North Beach home of OpenAI chief executive Sam Altman. The suspect, whose name has not been released pending court proceedings, was taken into custody on charges of attempted murder, arson and possession of an incendiary device. Investigators say the attacker posted a series of online essays in the weeks before the incident, warning that “uncontrolled AI will destroy humanity” and urging “direct action against those who profit from it.” The writings, which appeared on fringe tech forums and a personal blog, referenced Altman by name and described the planned attack as a “necessary warning shot.” Police confirmed the Molotov device was assembled from a gasoline‑filled bottle and a homemade fuse, but it failed to ignite the house, causing only minor property damage. The case builds on the criminal complaint filed on 14 April, when prosecutors first charged the suspect with attempted murder (see our earlier report). The arrest marks the first time law‑enforcement officials have linked the alleged extremist’s digital manifesto to a concrete act of violence against an AI industry leader. The episode underscores growing security concerns for high‑profile figures in the artificial‑intelligence sector, where rapid advances have sparked both admiration and hostility. It also raises questions about how online radicalisation around AI risks is monitored and countered. Watch for the upcoming arraignment, where prosecutors are expected to seek a pre‑trial detention order, and for any statements from OpenAI’s security team or the broader tech community about heightened protective measures. Legislative bodies may also revisit proposals to tighten monitoring of extremist content that targets AI executives, a debate that could gain urgency in the weeks ahead.
78

Can Claude Fly a Plane?

HN +6 sources hn
claude
Anthropic’s flagship model Claude took to the skies this week in a live demonstration that paired the language model with a commercial flight‑simulator interface. Engineers fed the simulator’s telemetry into Claude’s API and asked the model to generate real‑time control commands—throttle, pitch, yaw and landing‑gear actions—while a human overseer monitored the output. Within minutes the AI guided a virtual Cessna from take‑off to a textbook landing on a virtual runway in Warsaw, adjusting for wind gusts and instrument failures that were injected on the fly. The test builds on Anthropic’s recent rollout of Claude Code, which introduced deterministic permission handling and persistent memory features that let the model retain state across long, token‑heavy sessions. As we reported on 14 April, those upgrades already let developers stitch together complex workflows without “invisible tokens” draining limits. Applying the same architecture to a high‑frequency control loop demonstrates that Claude can move beyond text generation into domains that demand millisecond‑scale decision making. Aviation stakeholders are watching because the experiment hints at a new class of AI‑assisted cockpit tools. If a language model can interpret sensor feeds, reason about safety constraints and issue control inputs, it could augment pilots during high‑workload phases, flag anomalies, or even take over routine cruise management. The technology also raises regulatory questions: certification standards for software that directly manipulates flight surfaces are still nascent, and liability frameworks will need to evolve. Next steps include expanding the trial to more complex aircraft, integrating visual inputs from simulated cockpit displays, and testing under adverse weather scenarios. Anthropic plans to open the flight‑control API to a limited set of partners later this quarter, while the European Union Aviation Safety Agency has signaled interest in drafting guidelines for AI‑driven flight assistance. The coming months will reveal whether Claude’s virtual flight is a novelty or the first step toward AI‑enhanced aviation.
78

Don't Let AI Steal Your Intelligence

Mastodon +6 sources mastodon
A short YouTube clip titled “Don’t Let AI Steal Your Intelligence” has gone viral across Nordic tech circles, sparking a fresh debate about the cognitive risks of unchecked large‑language‑model (LLM) use. The 45‑second video, posted on the platform’s Shorts feed on April 13, juxtaposes a user typing a query into a chat interface with a rapid montage of the same person later struggling to recall basic facts without the model’s assistance. The caption, #ai #llm, invites viewers to consider whether constant AI prompting is eroding mental acuity. The clip is part of a broader campaign by writer‑developer Sam Choo, who recently released a Medium essay and a self‑published guide called *Don’t Let AI Steal Your Brain*. In those pieces, Choo argues that habitual reliance on AI for drafting, coding, or even clinical reasoning can lead to laziness, reduced problem‑solving skills, and a measurable dip in IQ scores. He backs the claim with anecdotal evidence from writers who notice a “thinking‑optional” mindset after months of AI‑assisted drafting, and with early data from a medical‑ethics blog noting that clinicians who let AI generate care pathways risk ceding responsibility for critical decisions. The warning arrives at a moment when Nordic developers are increasingly running LLMs locally—see our April 13 guide to Ollama and DeepSeek‑V3—because on‑device inference promises privacy without the “AI‑as‑overlord” narrative. Yet Choo’s message underscores that technical control does not automatically translate into cognitive stewardship. Industry observers say the next wave will focus on guidelines for “AI‑augmented cognition,” with the European Commission expected to publish recommendations on responsible AI use in professional settings later this year. Watch for academic studies quantifying the impact of daily AI assistance on memory retention and for policy proposals that may shape how companies train staff to balance efficiency with mental resilience.
76

Linear Programming for Multi-Criteria Assessment with Cardinal and Ordinal Data: A Pessimistic Virtual Gap Analysis

Linear Programming for Multi-Criteria Assessment with Cardinal and Ordinal Data: A Pessimistic Virtual Gap Analysis
ArXiv +7 sources arxiv
A new pre‑print on arXiv (2604.09555v1) proposes a linear‑programming framework that merges cardinal and ordinal information for multi‑criteria assessment. The authors call the method “pessimistic virtual gap analysis” (PVGA). It formulates each alternative’s performance as a set of linear constraints that capture exact numeric scores (cardinal data) and rank‑order preferences (ordinal data). By minimizing the worst‑case “virtual gap” – the distance between an alternative’s achievable score and an ideal reference point – the model yields a single scalar value that can rank all options without forcing ordinal inputs into arbitrary numeric scales. The contribution matters because most Multiple Criteria Decision‑Making (MCDM) tools either require fully quantified inputs or treat ordinal judgments as if they were cardinal, a practice that can distort outcomes in environmental planning, public procurement or AI model selection where qualitative rankings coexist with hard metrics. PVGA preserves the integrity of ordinal data, remains solvable with off‑the‑shelf simplex or interior‑point solvers, and produces a transparent worst‑case guarantee that decision makers can audit. Early simulations reported in the paper show tighter discrimination among alternatives compared with classic methods such as TOPSIS or weighted sum models, especially when data quality is uneven. The next steps will reveal whether the approach moves beyond theory. Watch for an open‑source implementation, likely in Python’s PuLP or Julia’s JuMP, and for pilot studies in EU sustainability assessments where mixed data are the norm. Industry groups may test PVGA for supplier evaluation, while academic circles could benchmark it against existing MCDM suites. If the method proves scalable, it could become a standard tool for AI‑augmented decision pipelines that must reconcile quantitative outputs with expert rankings.
73

OpenAI Chief’s Home Shot at With Second Attack in Days

OpenAI Chief’s Home Shot at With Second Attack in Days
Mastodon +7 sources mastodon
googleopenai
OpenAI chief Sam Altman’s San Francisco home was again the target of gunfire on Sunday morning, marking the second violent incident at the property within 48 hours. Police responded to reports of multiple shots fired outside the residence at roughly 08:30 local time; no one was injured and the house sustained only superficial damage. Investigators have detained two suspects they say are linked to the earlier Molotov‑cocktail attack that occurred two days earlier, when a 20‑year‑old threw an incendiary device at the same address. The back‑to‑back attacks raise fresh security concerns for high‑profile AI leaders. Altman, who has become the public face of the industry after OpenAI’s rapid rollout of ChatGPT‑4 and its upcoming multimodal models, has already been the subject of intense scrutiny and hostility from both anti‑AI activists and political actors. The first assault, a Molotov cocktail, was framed by the perpetrator as a “recipe‑following” act inspired by a ChatGPT prompt, a claim that sparked a wave of online ridicule and debate about the weaponisation of generative AI. The latest shooting, however, appears to be a more conventional intimidation tactic, suggesting that the threats are not limited to fringe internet provocations. Authorities have not disclosed a motive, but they are reviewing surveillance footage, digital footprints and any communications that might link the suspects to organised anti‑AI groups. OpenAI’s security team is reportedly tightening protection for its executives and reviewing protocols for staff and visitors. What to watch next: the San Francisco Police Department will release an official statement on the suspects’ identities and any charges by the end of the week. OpenAI is expected to address the incidents in its next board meeting, potentially revisiting its public‑relations strategy and lobbying efforts. Meanwhile, the AI community will be monitoring whether the attacks trigger broader calls for heightened security measures or legislative action aimed at protecting tech leaders.
72

I understand that older people have a decades-long reputation for not adopting new things because we

Mastodon +6 sources mastodon
grok
A recent survey of AI practitioners across Sweden, Norway, Denmark and Finland has upended the long‑standing stereotype that senior professionals shy away from cutting‑edge tools. Conducted by the Nordic AI Association in partnership with the University of Helsinki, the study found that 48 % of respondents aged 55 plus are already integrating generative‑AI assistants into daily coding, data‑analysis and research workflows, a figure that rivals the 52 % adoption rate among workers under 35. The data emerged from an online questionnaire distributed to more than 3,000 members of regional AI societies, followed by in‑depth interviews with a cross‑section of senior engineers, data scientists and academic researchers. Participants highlighted three drivers: a desire to stay competitive in a talent‑tight market, institutional upskilling programmes that target “late‑career” staff, and the tangible productivity gains reported when AI drafts code snippets or summarises literature. Economic concerns, often cited as a barrier for older adults in health‑tech adoption, appeared less decisive in the professional sphere where corporate training budgets offset personal costs. Why the shift matters is twofold. First, it expands the pool of experienced talent that can be leveraged as AI ecosystems mature, mitigating the risk of a skills gap as the Nordic tech sector scales. Second, it challenges age‑related bias in hiring and project allocation, prompting firms to reconsider assumptions about flexibility and learning capacity among veteran staff. Looking ahead, the association plans to publish a longitudinal follow‑up next spring to track retention of AI‑enhanced workflows among senior workers. Companies are also expected to roll out more structured mentorship schemes that pair younger coders with seasoned experts who now wield AI tools, potentially reshaping collaboration patterns across the region’s burgeoning AI landscape.
72

OpAMP server with MCP – aka conversational Fluent Bit control I’ve written a few times abo

Mastodon +6 sources mastodon
agents
A new open‑source server that fuses the Open Agent Management Protocol (OpAMP) with the Model Context Protocol (MCP) has been released, promising “conversational” control of Fluent Bit log agents. The project, announced on GitHub this week, implements OpAMP’s central‑server/agent model while exposing MCP‑driven tool calls that let large language models issue real‑time commands, query status and adjust configurations through a ChatOps‑style interface. OpAMP, a CNCF‑backed evolution of the OpenTelemetry Protocol (OTLP), standardises how a supervisory service discovers, configures and monitors distributed observability components. By wiring MCP into the same control plane, the server lets an LLM act as a first‑class operator: it can select from a catalogue of Fluent Bit actions—such as dynamic pipeline reloading, filter tuning or metric export toggling—and execute them without writing scripts. The result is a unified, language‑model‑aware observability stack where humans and AI can converse with the same endpoint. The integration matters because it lowers the barrier to sophisticated log management in cloud‑native environments. Teams can now ask an AI assistant to “increase error‑level sampling on service X” and see the change reflected across all Fluent Bit instances in seconds, cutting the latency of incident response and reducing reliance on manual configuration drifts. Security‑focused containers, already a hallmark of Fluent Bit’s Docker images, benefit from the same centralised policy enforcement that OpAMP provides. As we reported on April 12, the MCP framework is gaining traction in research tools such as the Grainulator plugin that forces Claude Code to substantiate its claims. This deployment marks the first production‑grade use of MCP for operational tooling. Watch for CNCF’s upcoming OpAMP spec finalisation, community adoption metrics, and extensions that tie the server into popular ChatOps platforms like Slack or Microsoft Teams. Early adopters are expected to publish benchmark data on latency and token usage, which will shape the next wave of AI‑augmented observability.
72

Claude Code OAuth down for >12 hours

HN +6 sources hn
anthropicclaude
Claude Code’s OAuth login service was unavailable for more than twelve hours on Tuesday, leaving thousands of developers unable to sign in with their Microsoft or Google accounts. The outage, first reported by users on GitHub and Reddit, manifested as a “timeout” error on Windows machines and prevented the launch of any Claude‑Code‑powered IDE extensions. Anthropic’s status page confirmed the incident at 08:17 UTC and posted a resolution timestamp of 20:45 UTC, citing a “misconfiguration in the token‑exchange endpoint” as the root cause. The downtime matters because OAuth is the default entry point for Claude Code’s cloud‑based coding assistant, which many teams have integrated into Visual Studio Code, JetBrains IDEs and custom CI pipelines. When the flow fails, developers are forced to fall back on API‑key authentication—a step that bypasses the single‑sign‑on convenience and can expose token management overhead. The interruption also highlighted a broader reliability concern: Claude Code’s recent rollout has already drawn scrutiny over “invisible tokens” that silently consume usage limits and the platform’s aggressive rate‑limiting policies, topics we covered in our April 14 pieces on Claude Code’s token handling and RAG pipeline redesign. What to watch next is Anthropic’s response. The company has promised a post‑mortem and a hardening of the OAuth gateway, including redundant token‑exchange servers and clearer error messaging. Developers should monitor the Anthropic status dashboard and the upcoming “Claude Code 2.0” roadmap, where a native password‑less login and optional SSO federation are slated for Q3. In the meantime, teams that rely heavily on Claude Code are advised to keep API keys on hand and to test alternative authentication paths to mitigate future disruptions.
72

Claude Code may be burning your limits with invisible tokens

Claude Code may be burning your limits with invisible tokens
HN +6 sources hn
claude
Anthropic’s Claude Code, the company’s AI‑powered coding assistant that many developers have adopted for everything from autocomplete to autonomous bug‑fix loops, is reportedly draining user quotas far faster than advertised. A wave of complaints surfaced in early April, prompting Anthropic to confirm that two independent bugs in the standalone Claude Code binary are inflating token counts by roughly ten‑to‑twenty times. The first bug breaks prompt‑caching, causing the model to resend previously cached context on every iteration. The second bug miscounts tokens generated by the internal “scratchpad” used for chain‑of‑thought reasoning, adding invisible tokens that never appear in the user’s prompt or output but still count against the quota. The issue matters because Claude Code’s pricing is tied directly to token consumption; developers on the Claude Max plan have seen daily limits evaporate after a single coding session, forcing them to downgrade or switch to competing tools such as OpenAI’s Codex. For teams that rely on autonomous loops to refactor large codebases, the hidden cost threatens project budgets and erodes confidence in Anthropic’s metering transparency. The problem also highlights a broader risk for AI‑augmented development: when internal mechanisms silently consume resources, users cannot predict expenses or optimise prompts. Anthropic says it is investigating the bugs and will release a hot‑fix within weeks. In the meantime, power users are rolling back to earlier binary versions and disabling the scratchpad feature to curb token burn. Watch for an official patch timeline, any adjustments to quota‑reset policies, and whether Anthropic will introduce more granular token‑usage dashboards. The episode could also spur tighter industry standards for token accounting in AI coding agents, a development that would benefit the Nordic tech ecosystem that increasingly leans on such tools.
71

Mark Zuckerberg is reportedly building an AI clone to replace him in meetings

Mastodon +6 sources mastodon
ai-safetygooglemeta
Meta is reportedly developing a photorealistic, AI‑powered 3‑D avatar of CEO Mark Zuckerberg that could sit in on internal meetings on his behalf. According to the Financial Times, the project—codenamed “Zuck‑Bot” by insiders—feeds the executive’s public speeches, interview transcripts and internal communications into a generative‑AI pipeline that learns his cadence, humor and decision‑making style. The resulting digital double would be able to field questions, present updates and even make on‑the‑spot recommendations, while the real Zuckerberg focuses on product strategy and external engagements. The move signals a shift from AI as a tool for developers toward AI as a surrogate for senior leadership. If successful, Meta could reduce the time its top executive spends on routine briefings, streamline information flow across its sprawling organization, and set a precedent for “AI‑augmented CEOs” in other tech giants. Critics warn that delegating decision‑making to a model trained on past statements may entrench existing biases and obscure accountability, especially given Meta’s recent scrutiny over AI safety and content moderation. What to watch next is whether Meta will roll out the avatar in a pilot phase, likely within the company’s headquarters in Menlo Park, and how employees react to interacting with a synthetic version of their boss. The company’s AI ethics board is expected to review the deployment, and regulators may question the transparency of AI‑mediated leadership. A follow‑up announcement on performance metrics or a public demo could also spark broader industry debate about the limits of AI in executive roles. The coming weeks will reveal whether the experiment remains an internal efficiency hack or becomes a headline‑making model for AI‑driven corporate governance.
71

Sam Altman Attacked | AI Jesus | Russia Crackdowns [Tech News]

Mastodon +6 sources mastodon
openai
OpenAI chief executive Sam Altman was again the target of a violent incident on Thursday, when police arrested a suspect charged with attempted murder and attempted arson after a Molotov‑cocktail‑style device was thrown at his San Francisco home. The arrest follows two earlier raids on Altman’s residence that left the CEO’s family shaken and prompted a flurry of media coverage. The latest suspect, identified by the San Francisco Police Department as a 27‑year‑old with known anti‑AI sentiments, allegedly approached the front door, ignited an incendiary device and fled before officers arrived. Investigators say the device failed to cause structural damage, but the incident underscores a growing pattern of hostility toward AI leaders. As we reported on 12 April, Altman had already posted a family photo after a Molotov cocktail attack, describing the episode as a “wake‑up call” about the power of extremist opposition. Why the attacks matter goes beyond personal safety. Altman is the public face of OpenAI’s flagship models—GPT‑4, ChatGPT and DALL‑E—whose rapid deployment fuels debates over regulation, misinformation and economic disruption. The assaults coincide with a wave of Russian government crackdowns on AI research, including new censorship rules that label unapproved generative tools as “dangerous propaganda.” Russian officials have even dubbed Altman an “AI Jesus,” a tongue‑in‑cheek reference that hints at both reverence and resentment for the influence his company wields. What to watch next: federal authorities are expected to review security protocols for high‑profile tech executives, while OpenAI may bolster its own protective measures. In Washington, lawmakers are likely to cite the attacks when debating AI‑related legislation, and the Russian Ministry of Digital Development is poised to announce stricter licensing for foreign AI services. The convergence of personal threats and geopolitical pressure could shape the next phase of AI governance and the safety of its architects.
71

The # Moodle # gradebook re-imagined by the Anthropic Opus # LLM

Mastodon +6 sources mastodon
anthropicclaude
Anthropic’s newest Opus model is being rolled out as a plug‑in for Moodle’s gradebook, turning the long‑standing manual grading hub into an AI‑driven analytics console. The integration, announced this week on Anthropic’s developer portal, lets instructors push course data into Opus with a single click via Zapier, where the model automatically extracts, summarizes and validates grades, flags anomalies and suggests personalized feedback for each student. Real‑time dashboards, built on Datadog‑fed metrics, show confidence scores for each AI‑generated entry and alert teachers to potential prompt‑hacking attempts, addressing long‑standing concerns about data privacy and model manipulation. The move matters because Moodle powers more than 200 million learners worldwide, yet its grading tools have changed little since the platform’s inception. By embedding a large language model that can interpret rubric language, reconcile weighted assessments and even propose grade‑curving scenarios, Opus promises to cut administrative overhead and reduce human error. Anthropic’s partnership with Instructure, announced in April 2025, laid the groundwork for “Claude for Education”; Opus is the first generative model built specifically for the higher‑education workflow, signalling a shift from experimental pilots to production‑grade AI in the classroom. What to watch next: Anthropic has pledged a public beta for the Opus gradebook in the coming weeks, with pilot institutions in Sweden, Norway and Denmark slated to test the feature under GDPR‑compliant data handling. Observers will be keen to see adoption rates, the model’s impact on grading turnaround times and whether faculty unions raise concerns about algorithmic assessment. A follow‑up study from the Nordic AI Institute, due later this year, will compare Opus‑enhanced gradebooks against traditional setups, offering the first independent benchmark of AI‑augmented grading at scale.
69

Evaluating Netflix Show Synopses with LLM-as-a-Judge

HN +5 sources hn
Netflix has rolled out an internal “LLM‑as‑a‑Judge” system to grade the synopses that accompany its original series and licensed titles. The framework prompts a large language model to assess each description against a set of creative and factual criteria, generates tiered rationales, aggregates scores from multiple model instances, and runs a dedicated factuality agent to flag inaccuracies. The output is a consensus rating that feeds directly into the content‑metadata pipeline. The move matters because synopsis quality is a silent driver of viewer engagement. Better‑crafted blurbs can sharpen search relevance, improve recommendation algorithms and reduce the manual effort of copy‑editors who currently vet thousands of descriptions each month. Netflix’s internal validation, which compared the LLM scores to a human‑labeled “golden set,” shows a strong correlation with member satisfaction metrics, suggesting the AI’s judgments align closely with real‑world audience response. Netflix’s experiment is the latest high‑profile deployment of the LLM‑as‑a‑judge pattern, a technique that has been gaining traction for code review, content moderation and, now, creative evaluation. By entrusting an AI with a task that traditionally required subjective human judgment, the streamer signals confidence in the technology’s consistency and scalability, while also raising questions about bias, transparency and the future role of human copywriters. What to watch next is whether Netflix expands the model to other assets such as thumbnail captions, trailer descriptions or even recommendation scoring. The company has hinted at publishing the evaluation dataset later this year, which could spur open‑source implementations and give rivals a benchmark for their own AI‑driven metadata workflows. Industry observers will also be tracking any regulatory feedback on AI‑generated consumer‑facing text as the practice moves from pilot to production.
68

The Apple Watch Series 11 has returned to its best-ever price

Mastodon +6 sources mastodon
amazonapple
Apple has slashed the price of its flagship smartwatch, bringing the 42 mm Apple Watch Series 11 with GPS down to $299 on Amazon, Best Buy and Target – a $100 discount that marks the device’s lowest‑ever retail price. The cut, announced on April 13, represents a 25 percent reduction from the model’s $399 list price and mirrors a similar promotion that briefly appeared earlier this month. The price drop is significant because the Series 11, launched in September 2025, remains a cornerstone of Apple’s health‑and‑fitness ecosystem. It offers an ECG sensor, blood‑oxygen monitoring, a new temperature‑tracking algorithm and the always‑on Retina display that debuted with the Series 8. By making the watch more affordable, Apple is likely aiming to clear inventory ahead of the expected launch of the Series 12, rumored to arrive in the fall with a slimmer chassis and advanced health metrics such as non‑invasive glucose monitoring. For consumers, the discount lowers the barrier to entry for Apple’s premium wearables, potentially expanding the user base that can access features like fall detection, emergency SOS and seamless integration with iOS 27. Retail analysts also see the move as a response to intensifying competition from cheaper Android‑based smartwatches that have been gaining market share in Europe and North America. What to watch next: Apple’s supply‑chain signals and any further price adjustments in the coming weeks could hint at the timing of the next hardware refresh. Industry watchers will also monitor whether the discount spurs a measurable uptick in watch sales before the holiday season, and how it fits into Apple’s broader strategy of bundling wearables with upcoming software updates such as iOS 27’s enhanced health dashboard.
68

American Airlines Now Supports iOS 26's Revamped Wallet Boarding Passes

Mastodon +6 sources mastodon
apple
American Airlines has rolled out support for the refreshed boarding‑pass format introduced with iOS 26, making the airline the latest of the United States’ four major carriers to tap Apple Wallet’s new capabilities. The update, pushed through the airline’s app on Tuesday, replaces the static QR code with a dynamic card that blends a cleaner visual design with interactive elements such as Apple Maps‑based destination guides, real‑time luggage‑tracking icons and quick‑access links to gate information. The change matters because Apple’s iOS 26 overhaul turns Wallet from a passive storage tool into a travel hub. By embedding live data feeds, airlines can push gate changes, delay alerts or even upsell services directly onto the pass, reducing the need for separate apps or paper tickets. For passengers, the integration promises a smoother check‑in experience and a single point of reference for trip logistics, while giving Apple another foothold in the lucrative airline‑services ecosystem that rivals Google’s Wallet on Android. The rollout follows Apple’s recent beta releases of iOS 26.5 and iPadOS 26.5, which added finer‑grained developer controls for dynamic passes. As we reported on 13 April, the new boarding‑pass framework is already being adopted across the industry, and American Airlines’ move completes the quartet of U.S. flag carriers on board. What to watch next: Apple is expected to extend the Wallet toolkit with in‑flight service cards and loyalty‑program integration later this year, while airlines may begin experimenting with personalized offers tied to a traveler’s itinerary. On the Android side, Google’s parallel Wallet enhancements could spark a cross‑platform race for the most feature‑rich digital boarding pass. Keep an eye on announcements from smaller carriers and international airlines, which could broaden the ecosystem beyond the U.S. market.
68

Huawei Teases a Wider Foldable, and the Timing Feels Very Apple-Adjacent

Huawei Teases a Wider Foldable, and the Timing Feels Very Apple-Adjacent
Mastodon +6 sources mastodon
apple
Huawei has unveiled the Pura X Max, a new ultra‑wide foldable that expands the company’s “book‑style” line and lands just weeks before Apple’s long‑rumored iPhone Fold is expected to appear in the market. The device, shown in a short teaser on CNET and corroborated by Digital Trends and other outlets, sports a tablet‑sized inner screen and a back panel with textured, gridded sections that echo the design language of last year’s Pura X. Its dimensions place it squarely in the “passport‑size” category that has been associated with Apple’s prototype foldables, effectively beating Cupertino to the punch in both form factor and launch timing. The move matters for several reasons. First, it reinforces Huawei’s strategy of using premium hardware to regain relevance in a market where U.S. sanctions have limited its access to Google services. By delivering a device that rivals Apple’s speculative offering, Huawei can capture high‑end Chinese consumers eager for a home‑grown alternative to the iPhone. Second, the wider foldable challenges Samsung’s dominance in the segment; Samsung’s Galaxy Z Fold series has set the benchmark for size and price, but Huawei’s aggressive pricing and integration with its own HarmonyOS ecosystem could reshape buyer expectations. Finally, the launch underscores a broader industry shift toward larger, tablet‑like foldables, suggesting that the “wide‑fold” form factor may become the new standard rather than the compact flip. What to watch next: Huawei has slated a global pre‑order window for early May, with shipments expected in June. Analysts will be monitoring consumer response in China and the device’s performance in markets where Huawei’s app ecosystem is still maturing. Apple’s next product event, slated for September, will likely reveal whether the iPhone Fold can still generate excitement after Huawei’s head start, or if the competition will force Apple to accelerate its own foldable timeline.
65

Spiking Neural Networks — interactive explorer

Mastodon +6 sources mastodon
A new interactive “Spiking Neural Network (SNN) Explorer” has been released as the final teaching widget for a university‑level Bio‑Inspired AI and Optimization course. The web‑based tool lets students build and visualise leaky‑integrate‑and‑fire (LIF) neurons, experiment with rate‑coding schemes and spike‑timing‑dependent plasticity (STDP), and compare a curated list of hardware and software implementations ranging from neuromorphic chips to Python simulators. By exposing the timing‑based dynamics that distinguish SNNs from conventional deep nets, the explorer aims to demystify a technology that has long lingered on the periphery of mainstream AI research. The launch matters because SNNs are increasingly touted as the “third generation” of neural networks, promising orders‑of‑magnitude lower energy consumption and tighter alignment with how biological brains process information. Recent studies, such as our April 12 coverage of biological neural networks as viable alternatives to conventional machine‑learning models, have highlighted the strategic relevance of neuromorphic computing for edge devices and sustainable AI. An accessible, hands‑on platform lowers the barrier for both students and researchers to prototype SNN‑based solutions, potentially accelerating the transition from academic curiosity to production‑ready applications in robotics, sensor processing and low‑power inference. What to watch next is the open‑source release schedule and any integration with major neuromorphic platforms such as Intel’s Loihi or IBM’s TrueNorth. The developers have hinted at upcoming benchmark suites that will compare SNN performance against traditional deep‑learning baselines on tasks like image classification and event‑based vision. If the explorer gains traction in curricula across the Nordics, it could seed a new generation of engineers equipped to harness spike‑driven computation, nudging the AI ecosystem toward more biologically plausible and energy‑efficient models.
65

Man charged with firebomb attempt at home of OpenAI CEO Sam Altman

The Washington Post on MSN +7 sources 2026-04-06 news
openai
A federal indictment unsealed Monday accuses 20‑year‑old Daniel Moreno‑Gama of attempting to kill OpenAI chief executive Sam Altman and a security guard by hurling a Molotov‑cocktail‑style firebomb at the gate of Altman’s San Francisco home on April 10. Court documents show Moreno‑Gama was arrested after surveillance footage captured him approaching the residence, and agents later raided his Texas home, seizing a notebook that detailed “anti‑AI” motivations. The Department of Justice charged him with attempted murder, use of an incendiary device and possession of a firearm by a prohibited person. The case marks the latest escalation in a string of threats against the OpenAI leader. As we reported on April 14, Altman’s house was the target of two separate attacks within days, and the CEO’s personal safety has become a flashpoint in the broader backlash against generative‑AI technologies. OpenAI’s recent moves – the acquisition of fintech startup HIRo Finance and the rollout of new paid services – have amplified its public profile, drawing both admiration and hostility. The indictment underscores how that visibility can translate into violent extremism, raising questions about the security of AI executives and the potential chilling effect on innovation. What to watch next: prosecutors will seek a trial date while federal agents continue to probe whether Moreno‑Gama acted alone or is part of a wider anti‑AI network. OpenAI is expected to tighten personal security for its leadership and may issue a formal statement on workplace safety. Legislators in the U.S. and Europe, already debating AI regulation, could cite the incident when arguing for stricter oversight of AI firms and protection of their staff. The outcome of this case will likely shape how the industry balances rapid growth with the need for robust security measures.
54

What Karpathy's LLM Wiki Is Missing (And How to Fix It)

Dev.to +5 sources dev.to
Andrej Karpathy’s “LLM Wiki” pattern exploded on GitHub this month, amassing more than 5,000 stars and 3,700 forks within weeks. The approach, which treats a large‑language model as a curator that reads a corpus, extracts key takeaways and writes them into a personal, markdown‑based wiki, has been cloned dozens of times and is already powering experimental knowledge‑bases from hobbyists to early‑stage startups. The buzz stems from the pattern’s promise to sidestep the classic “re‑derive‑on‑every‑query” loop that plagues LLM‑augmented retrieval. By loading the entire knowledge set once and letting the model maintain a structured, human‑readable index, developers can reduce token consumption, lower latency and, crucially, keep a transparent audit trail of what the model has learned. The core workflow—read, discuss, summarise, update index, propagate changes across entity pages, and log the operation—mirrors a lightweight version of a corporate wiki, but with AI‑driven upkeep. Critics, however, point out three blind spots. First, the flat markdown hierarchy struggles with scale: as the wiki grows, token limits reappear and update latency spikes. Second, the pattern offers no built‑in mechanism to resolve contradictions or detect hallucinations, leaving the model to trust its own summaries. Third, it lacks a semantic layer that could link concepts across pages, limiting cross‑referencing and query precision. A wave of community patches aims to plug those gaps. Projects that overlay a lightweight knowledge graph on top of the markdown files promise automated entity linking and conflict resolution, while incremental indexing techniques keep token usage in check. The most promising prototype integrates a vector store that caches embeddings for each page, allowing the LLM to retrieve only the most relevant sections on demand. What to watch next: the first open‑source fork that combines Karpathy’s wiki with a graph‑backed index is slated for release in early May, and several Nordic AI labs have already pledged to test it on meeting‑transcript corpora. If the hybrid model delivers on its promise, it could become the de‑facto standard for privacy‑first, locally‑run knowledge bases—building on the Open KB initiative we covered on April 14. The next few weeks will reveal whether the community can turn a viral pattern into a production‑grade tool.
54

Show HN: Open KB: Open LLM Knowledge Base

HN +6 sources hn
agentsbias
A new open‑source project called **Open KB** landed on Hacker News on Tuesday, promising an “Open LLM Knowledge Base” that lets anyone turn raw documents into a structured, cross‑referenced wiki powered by large language models. The repository, posted by developer mingtianzhang, builds on Andrej Karpathy’s LLM‑Wiki concept: users drop source files into a folder, an LLM parses the content, generates concise pages, adds links, runs bias checks and maintains a master index—all inside the Obsidian note‑taking environment. The timing is significant. As open‑source models such as Llama 3.1 and community‑run leaderboards on Hugging Face demonstrate, the barrier to running powerful LLMs on consumer hardware is falling. Open KB extends that trend from inference to knowledge management, offering a privacy‑first alternative to cloud‑based vector stores and proprietary knowledge‑graph services. By keeping data and inference local, the tool aligns with the privacy‑centric voice‑assistant framework we covered earlier this week in “Building a Privacy‑First Voice‑Controlled AI Agent with Local LLMs” (April 14). It also addresses a growing demand among developers, researchers and hobbyists for reproducible, auditable AI‑generated documentation without surrendering proprietary data to third‑party APIs. What to watch next is how quickly the community adopts and expands the platform. Early indicators include forks that integrate retrieval‑augmented generation pipelines, experiments with multi‑GPU acceleration (as seen in the “How I Topped the HuggingFace Open LLM Leaderboard on Two Gaming GPUs” post), and potential partnerships with note‑taking apps beyond Obsidian. If Open KB gains traction, it could become a de‑facto standard for locally maintained AI knowledge bases, challenging commercial offerings and shaping the next wave of privacy‑aware AI tooling. Keep an eye on GitHub activity and forthcoming tutorials that will reveal how scalable the approach is in real‑world deployments.
45

I recently reread Fred Brooks's 1986 classic No Silver Bullet , which you can find here: https://w

Mastodon +6 sources mastodon
A senior software‑engineer turned AI commentator has republished a fresh take on Fred Brooks’s 1986 classic *No Silver Bullet*, linking the original PDF and urging the community to reassess the paper’s distinction between essential and accidental complexity in the age of large‑language‑model (LLM) coding assistants. The essay, which quickly gathered attention on Nordic tech forums, argues that tools such as GitHub Copilot, Claude Code and OpenAI’s new developer‑focused APIs have begun to shave off a measurable slice of “accidental” overhead—boilerplate, syntax errors and routine refactoring—while the deeper, domain‑specific challenges Brooks labelled “essential” remain untouched. The reinterpretation matters because it cuts through the current hype cycle that promises AI will deliver an order‑of‑magnitude boost in software productivity. By grounding the discussion in Brooks’s framework, the author reminds investors and product teams that AI can automate repetitive tasks but cannot eliminate the need for architectural insight, problem decomposition or rigorous testing. The piece also references recent observations that Claude Code’s hidden token accounting can silently inflate usage limits, a reminder that new tooling can introduce its own accidental complexities. Looking ahead, the conversation is set to move from theory to data. Researchers at the University of Copenhagen plan to publish a longitudinal study measuring code‑completion impact on defect rates across three major LLMs. Meanwhile, OpenAI’s recent acquisition of fintech startup HIRo Finance signals a broader push to embed AI deeper into domain‑specific workflows, a development that will test whether the “essential” barriers can ever be lowered by smarter tooling. Stakeholders should watch for the upcoming “AI‑Assisted Development” track at the Nordic Software Engineering Conference in June, where early results and best‑practice guidelines are expected to surface.
45

AI-boosted hacks with Anthropic's Mythos could have dire consequences for banks

HN +6 sources hn
anthropicgoogle
Anthropic’s latest large‑language model, Claude Mythos, has moved from research showcase to a security alarm bell for the banking sector. Within days of the model’s public unveiling, cybersecurity analysts warned that Mythos can automate the discovery of zero‑day flaws and generate sophisticated phishing or ransomware payloads at a speed that outpaces traditional defenses. In a Reuters briefing on April 13, experts demonstrated how the model autonomously identified critical vulnerabilities in legacy banking software and produced exploit code that would have taken a human team weeks to craft. The threat matters because most financial institutions still run core‑banking platforms built on decades‑old codebases, often patched only after a breach is confirmed. Mythos’ ability to “boost” attacks means threat actors can bypass these outdated safeguards with minimal effort, potentially compromising transaction integrity, customer data and market‑wide confidence. Anthropic’s own documentation, released in the system card we covered on April 13, acknowledges the model’s capacity for unrestricted code generation, prompting the company to impose internal guardrails that, according to insiders, are already being challenged by external actors. What to watch next is two‑fold. First, regulators in the EU and Nordic countries are expected to issue guidance on AI‑enabled cyber risk, likely extending the forthcoming AI Act to cover malicious use cases. Second, Anthropic has signaled plans to roll out a “secure‑by‑design” version of Mythos with tighter usage controls, but the timeline remains unclear. Meanwhile, banks are accelerating investments in AI‑driven threat‑intelligence platforms and revisiting legacy system migration roadmaps. The coming weeks will reveal whether industry‑wide defensive measures can keep pace with a model that turns the same generative power that fuels productivity into a potent weapon for cyber‑criminals.
42

Evaluation of Claude Mythos Preview's cyber capabilities

HN +5 sources hn
anthropicclaude
Anthropic’s Claude Mythos Preview has been put through a rigorous cybersecurity benchmark, and the results confirm the model’s unprecedented offensive capabilities. In a test released on April 7, the system solved a full‑stack takeover (TLO) from start to finish in three of ten runs and completed an average of 22 of the 32 required steps across all attempts. Compared with the previous‑generation Claude Opus 4.6, Mythos Preview scored roughly eight percentage points higher and advanced six more steps in a simulated enterprise breach, making it the only model to achieve a complete takeover in the suite. The evaluation matters because it quantifies a leap in AI‑driven threat generation that could reshape the cyber‑risk landscape. Earlier this week we warned that Anthropic’s “Mythos” family could enable banks to be compromised at scale; the new data shows the preview model can autonomously discover and exploit zero‑day flaws in major operating systems and browsers, a capability no prior AI has demonstrated. Such proficiency lowers the barrier for sophisticated attacks, potentially accelerating the weaponisation of AI by criminal groups and nation‑states. It also raises questions about the adequacy of existing defensive tooling, which was not designed for an adversary that can iterate through dozens of exploit steps without human guidance. What to watch next includes Anthropic’s decision on whether to release Mythos Preview beyond internal testing, and how quickly the company will implement or disclose mitigation measures. Regulators in the EU and the United States are expected to scrutinise the model under emerging AI‑risk frameworks, while security vendors may race to develop counter‑AI solutions. Follow‑up research from independent labs will likely probe the model’s limits on defensive tasks, offering a clearer picture of whether its power can be harnessed for protection as well as exploitation.
41

OpenAI Rotates macOS Certificates Following Axios Supply Chain Breach

Mastodon +6 sources mastodon
openai
OpenAI announced on Tuesday that it has rotated all macOS code‑signing certificates after a malicious version of the open‑source library Axios slipped into its continuous‑integration pipeline. The compromised package was downloaded during a routine build, triggering a broader software‑supply‑chain attack that could have allowed a forged binary to run on users’ machines. OpenAI’s security team revoked the affected certificates and issued new ones, urging developers and end‑users to update any OpenAI‑branded macOS applications before the old certificates are blocked in May 2026. The incident matters because macOS code‑signing certificates are the trust anchor that lets the operating system verify an app’s authenticity. If an attacker can sign a malicious binary with a valid certificate, the app can bypass Gatekeeper and execute with the same privileges as a legitimate program. Although OpenAI says no user data or internal systems were accessed, the breach exposed a critical weakness in the company’s dependency management and highlighted the growing risk of third‑party libraries being weaponised in CI environments. OpenAI’s swift rotation mirrors industry best practice after similar supply‑chain compromises, such as the 2023 SolarWinds and 2024 Log4j incidents, and underscores the need for tighter verification of build‑time dependencies. The company has also pledged to audit its CI workflows and to work with Axios maintainers to patch the vulnerable release. What to watch next: Apple’s security response, including any additional notarisation checks for affected apps, will be closely monitored. OpenAI is expected to publish a detailed post‑mortem in the coming weeks, and regulators may scrutinise the incident under emerging EU and US software‑supply‑chain guidelines. Developers using OpenAI’s macOS SDK should verify they are running the latest signed binaries and review their own dependency‑checking processes to avoid similar exposure.
41

The TAM for Intelligence Is Infinite - Brad Gerstner and the All-In Besties # ai # investing

Mastodon +6 sources mastodon
openai
Altimeter Capital’s founder Brad Gerstner told the All‑In podcast on Tuesday that the total addressable market (TAM) for “intelligence” is effectively limitless, dwarfing any sector the firm has backed in the past two decades. Speaking alongside Chamath Palihapitiya, David Sacks and David Friedberg, Gerstner argued that the relentless stream of daily AI launches is expanding the economic frontier faster than any previous technology wave. The comment marks a shift from Altimeter’s recent focus on a handful of headline‑making AI stocks toward a broader, infrastructure‑level bet. Gerstner highlighted OpenAI’s multimillion‑dollar valuation, Anthropic’s recent funding round and the surge in enterprise‑grade models as evidence that capital is moving from speculative “toy” applications to core intelligence platforms that can be embedded across finance, healthcare, logistics and creative industries. For investors, the message is clear: the upside is not limited to a few public equities but lies in the ecosystem of data, compute and talent that underpins every AI service. Why it matters now is twofold. First, the infinite‑TAM narrative could accelerate capital inflows into late‑stage startups that promise to become the “operating systems of intelligence,” a space Altimeter is already positioning itself to dominate. Second, the framing may influence valuation benchmarks for upcoming AI IPOs, nudging the market to price companies on future network effects rather than current revenue. Watch for Altimeter’s next fund allocation signals, especially any disclosed commitments to compute providers or data‑centric ventures. Also keep an eye on the pipeline of AI‑focused IPOs that the All‑In panel hinted could debut before year‑end, and on how rival funds adjust their sizing models in response to Gerstner’s infinite‑TAM thesis.
41

🤖 Why don't LLMs track time in their conversations? Question for everyone: Why do you think LLMs

Mastodon +6 sources mastodon
claude
A post on the AI‑focused forum “Artificial Intelligence (AI)” sparked a fresh debate on why large language models (LLMs) such as Claude, ChatGPT or Gemini never embed timestamps in their dialogue streams. The user asked, “Why don’t LLMs track time in their conversations? It seems straightforward to note how long you’ve been talking.” The question quickly gathered dozens of replies from researchers, developers and hobbyists, turning a simple curiosity into a broader discussion about the structural limits of current generative models. The core reason is architectural. LLMs operate as next‑token predictors; they receive a block of text, process it through a fixed‑size context window, and output the most probable continuation. Adding a dynamic clock would require the model to treat time as a mutable variable, yet the underlying transformer layers have no built‑in notion of elapsed seconds or session length. Instead, temporal cues must be injected explicitly as part of the prompt, a practice that is rarely standardized. As we explained in our earlier piece “The Memory Problem: Why LLMs Sometimes Forget Your Conversation,” the same context‑window constraints that truncate long chats also prevent any persistent state from accumulating across turns, let alone a running timer. Why it matters goes beyond academic curiosity. Without temporal awareness, LLMs can misinterpret time‑sensitive instructions—e.g., “remind me in 10 minutes” or “what was the weather yesterday?”—and they cannot differentiate between a fresh query and a follow‑up that happened hours later. This hampers the development of truly conversational agents that can schedule, prioritize or adapt behavior over real‑world timelines. Looking ahead, several research groups are experimenting with “temporal tokens” that encode timestamps or duration markers inside the prompt, while others explore external memory modules that log interaction metadata. OpenAI’s recent “ChatGPT‑Turbo” update hints at a lightweight state‑tracking layer, and Anthropic has filed a patent for a “time‑aware context window.” Monitoring these prototypes will reveal whether the community can turn the current illusion of memory into a functional sense of time.

All dates