AI News

948

OpenAI shutting down Sora app

OpenAI shutting down Sora app
HN +9 sources hn
openaisora
OpenAI announced on Tuesday that it will discontinue Sora, the short‑form video‑generation app it launched six months ago. The company posted a brief note on X, thanking creators and confirming that both the consumer app and the professional‑grade internet service for film‑industry users will be shut down within weeks. Sora arrived with much fanfare as a TikTok‑style platform that let anyone produce AI‑generated clips from text prompts, showcasing OpenAI’s push into creative‑tool territory beyond text and image. Its rapid rise sparked excitement among content creators and concern among regulators, who warned that easy‑to‑make synthetic video could amplify deep‑fake risks and strain moderation resources. The shutdown, delivered without explanation, suggests the experiment ran into technical, safety or cost challenges that outweighed its early buzz. The move matters because it signals a recalibration of OpenAI’s product strategy. After a costly rollout of ChatGPT plugins and a surge in enterprise deals, the firm appears to be tightening its focus on higher‑margin services and tighter safety controls. Dropping Sora also removes a high‑visibility showcase of OpenAI’s video‑generation capabilities, potentially ceding ground to rivals such as Google’s Gemini Video, Meta’s Make‑it‑Real and emerging startups that are courting the same creator market. What to watch next: whether OpenAI will repurpose Sora’s underlying model for integration into ChatGPT or its upcoming enterprise suite, and how the company will address the regulatory pressure surrounding AI‑generated video. Competitors are likely to highlight the gap left by Sora, accelerating their own launches. Keep an eye on OpenAI’s next product announcements and any policy statements that clarify its stance on consumer‑facing generative video tools.
478

OpenAI Plans to Discontinue Support for Sora

OpenAI Plans to Discontinue Support for Sora
HN +7 sources hn
openaisora
OpenAI announced that it will cease all support for Sora, its short‑form AI‑generated video platform, by the end of the quarter. The decision covers the consumer app, the developer‑focused API, and the video‑generation feature embedded in ChatGPT. Employees were told the move is final, with no plan to revive the service or migrate users to an alternative offering. The shutdown marks a sharp reversal after Sora’s high‑profile launch last year, which was billed as a TikTok‑style competitor that could create licensed clips from text prompts. The venture attracted a $1 billion commitment from Disney, but the entertainment giant withdrew its investment earlier this month as OpenAI signalled the pull‑back. The loss of Disney’s backing, combined with tepid adoption and mounting legal scrutiny over copyrighted content and deep‑fake concerns, appears to have tipped the scales. For developers, the loss of the Sora API eliminates a rare tool for rapid video prototyping, forcing them to revert to slower, traditional pipelines or seek alternatives from rivals such as Meta’s Make‑It‑Real or Google’s Gemini Video. OpenAI’s broader strategy now leans toward text‑centric models and the automated‑researcher project highlighted in our March 24 coverage, suggesting the company is consolidating resources around higher‑margin products. Watch for how OpenAI reallocates the budget earmarked for Sora, whether it will launch a more tightly controlled video feature inside ChatGPT, and how Disney reshapes its AI partnership roadmap. Industry observers will also track regulatory responses to AI‑generated media, as the Sora episode underscores the tension between creative automation and intellectual‑property safeguards. The next few weeks should reveal whether OpenAI’s retreat from consumer video is a temporary retreat or a permanent shift away from the format.
335

OpenAI stellt Sora ein: Das Ende von immer weiter und größer

OpenAI stellt Sora ein: Das Ende von immer weiter und größer
Mastodon +7 sources mastodon
openaisora
OpenAI announced on Tuesday that it is shutting down Sora, its experimental AI‑video generator, marking a decisive shift in the company’s strategy. The move ends a three‑year partnership with Disney that let users create scenes populated by up to two hundred AI‑generated characters, and it follows a wave of criticism over deep‑fake potential and “AI‑trash” content. The closure signals that OpenAI is moving away from consumer‑facing, showcase‑type products toward revenue‑generating services for developers and enterprises. Analyst Eva‑Maria Weiß writes that the company now wants its ecosystem to “pay the bills” by embedding generative‑video capabilities into the broader API suite rather than maintaining a standalone app. The decision also reflects mounting competitive pressure: rivals such as Google DeepMind and Meta are rolling out their own multimodal models, while regulators tighten scrutiny on synthetic media. For users, the shutdown means the Sora web interface and associated APIs will disappear by the end of the month, with no migration path to a paid tier. Existing projects will need to be exported or rebuilt using alternative tools. OpenAI has not disclosed whether the underlying model will be repurposed for internal use or integrated into upcoming releases, including the just‑launched GPT‑5, which promises tighter multimodal integration. What to watch next is whether OpenAI will bundle video generation into its enterprise offerings, how it prices the feature, and how quickly competitors fill the gap left by Sora. The company’s next public statements on API roadmaps, as well as any regulatory actions on synthetic media, will indicate whether the pivot will translate into sustainable growth or simply a retreat from an over‑ambitious experiment. As we reported on 25 March, the Sora shutdown is the latest chapter in OpenAI’s rapid re‑orientation toward a B2B‑first model.
334

OpenAI brutally shuts down Sora, the most expensive slop factory in AI history

OpenAI brutally shuts down Sora, the most expensive slop factory in AI history
Mastodon +7 sources mastodon
openaisora
OpenAI announced on Thursday that it is shutting down Sora, the high‑profile AI video‑generation service it launched in September 2025. The company said the platform’s “mass‑scale video production costs a fortune, and no viable path to profitability has emerged.” The decision also ends a three‑year, $1 billion partnership with Disney that granted access to more than 200 licensed characters. As we reported on 25 March, Disney had already begun withdrawing from the deal after mounting concerns over uncontrolled deepfakes and the flood of low‑quality, disposable content Sora churned out. The abrupt closure confirms that the venture was unsustainable both financially and reputationally. OpenAI explained that the compute resources devoted to Sora will be redirected to its core language and image models, where demand and revenue prospects are clearer. The shutdown matters because it signals the first major retreat from large‑scale generative video, a field many had expected to explode after OpenAI’s high‑visibility launch. It underscores the difficulty of turning raw compute power into a profitable service when the output can be weaponised or quickly loses value. Regulators in Europe and the United States have been watching the deep‑fake debate closely, and the Sora episode may accelerate calls for stricter oversight of synthetic media. What to watch next: OpenAI’s next‑generation model roadmap may still include video capabilities, but likely as a research‑only feature rather than a consumer product. Disney is expected to announce alternative content‑creation strategies, possibly leveraging its own in‑house AI tools or partnering with smaller, niche providers. Finally, industry analysts will monitor whether other AI firms attempt a scaled‑video offering and how they address the cost‑and‑misuse challenges that doomed Sora.
318

OpenAI shutters Sora after it struggles to not be a copyright infringing mess! Is this the beginning

OpenAI shutters Sora after it struggles to not be a copyright infringing mess! Is this the beginning
Mastodon +8 sources mastodon
copyrightopenaisora
OpenAI announced on Tuesday that it is permanently disabling Sora, its short‑form AI video generator, just three months after signing a multiyear partnership with Disney to feature the studio’s characters. The company posted a brief note on X: “We’re saying goodbye to Sora.” In the same breath it confirmed that Disney has withdrawn from the $1 billion investment deal that underpinned the launch, citing “unresolved copyright‑risk concerns” and the app’s “inability to reliably filter infringing content.” The shutdown marks the swiftest reversal of a high‑profile product in OpenAI’s history. Sora debuted six months ago with a viral showcase of AI‑crafted clips, positioning the firm as a potential leader in automated video creation. Yet the technology struggled to meet the legal standards demanded by content owners, prompting a wave of takedown notices and a growing chorus of critics who warned that the model could become a “copyright‑infringing mess.” By pulling the plug, OpenAI is not only cutting a costly, under‑performing line but also signaling a broader retreat from experimental media tools in favour of its core text‑based models, enterprise APIs and the soon‑to‑be‑released GPT‑5. Why it matters extends beyond a single app. The episode underscores the regulatory and commercial headwinds facing generative‑AI firms that touch protected media, and it fuels speculation that the current AI hype cycle may be cooling. Investors and developers will be watching how OpenAI reallocates resources, whether it will revive video generation under stricter safeguards, and how rivals such as Google DeepMind or Meta respond with their own content‑aware tools. Next steps to monitor include OpenAI’s rollout of GPT‑5 later this year, any revised partnership strategy with entertainment studios, and potential policy actions from EU and US regulators aimed at curbing AI‑driven copyright violations. The Sora closure may prove a bellwether for how quickly the sector adapts to those pressures.
312

Local LLM App by Ente

Local LLM App by Ente
HN +6 sources hn
privacy
Ente has launched Ensu, a consumer‑grade app that runs a large language model entirely on the user’s device. The first version ships for macOS and iOS, with an Android beta slated for later this quarter. Ensu bundles a compact transformer model—optimised for Apple’s Neural Engine and Qualcomm’s Hexagon DSP—behind a sleek chat interface, while keeping all prompts and responses on‑device. Users can also enable a “Remote Tunnel” feature that forwards the model’s inference to a personal cloudflare‑hosted endpoint, letting them offload heavy workloads without exposing data to third‑party APIs. The release marks a tangible shift from the cloud‑centric AI services that dominate the market. By keeping the model local, Ente promises zero‑knowledge privacy, lower latency, and the ability to operate offline—attributes that appeal to privacy‑conscious consumers and enterprises wary of data‑leak risks. The move also underscores the rapid maturation of model compression techniques; a 7‑billion‑parameter model that once required a server‑grade GPU now fits within a smartphone’s memory budget. This follows our earlier coverage of hobbyists building private, local AI tools in a weekend with Ol, showing that the barrier to entry is collapsing from specialist to mainstream. What to watch next is the ecosystem that will grow around Ensu. Ente has opened a developer portal for plug‑ins, hinting at third‑party extensions such as domain‑specific knowledge bases or custom voice assistants. Analysts will be tracking adoption metrics on the App Store and any partnership announcements with hardware vendors that could embed the engine deeper into devices. A follow‑up update is expected in June when Ente plans to roll out a larger 13‑billion‑parameter model and expand support to Windows laptops, potentially setting a new baseline for on‑device AI performance.
297

OpenAI cancela Sora y se cae su acuerdo millonario con Disney En un giro inesperado, OpenA

OpenAI cancela Sora y se cae su acuerdo millonario con Disney      En un giro inesperado, OpenA
Mastodon +8 sources mastodon
openaisora
OpenAI announced on Tuesday that it is shutting down Sora, its generative‑video platform, and that the multimillion‑dollar partnership it had forged with Disney is now void. The company said the decision is final and that all Sora services will be taken offline within weeks, ending the short‑lived experiment that began with a high‑profile launch last year. The move marks a sharp reversal of OpenAI’s earlier push into consumer‑facing video AI. Sora, billed as a “TikTok‑style” app where users could upload prompts and receive AI‑generated clips, quickly attracted attention from creators and Hollywood alike, prompting Disney to sign a strategic alliance that promised co‑development of AI‑enhanced content and exclusive distribution rights. The partnership was seen as a test case for how legacy media could leverage large‑scale generative models to produce fresh IP at lower cost. OpenAI’s retreat matters for several reasons. First, it signals a shift away from high‑risk, low‑margin consumer products toward more profitable enterprise offerings such as ChatGPT Enterprise and custom‑model licensing. Second, the loss of Disney’s backing removes a marquee endorsement that could have accelerated industry adoption of AI video tools. Third, the shutdown raises questions about the viability of AI‑generated media under existing copyright and deep‑fake regulations, issues that have already drawn scrutiny from European and U.S. policymakers. What to watch next: OpenAI is expected to double down on its core language and image models, potentially accelerating the rollout of next‑generation GPT‑5. Disney, meanwhile, has hinted at pursuing its own in‑house AI capabilities, suggesting a possible partnership with a rival vendor. Analysts will also monitor whether other media giants, such as Warner Bros. Discovery or Netflix, step in to fill the vacuum left by the aborted Disney‑OpenAI collaboration. As we reported on March 25, OpenAI had already signaled the end of Sora; today’s announcement confirms that the venture is completely off the table.
280

What Sora's Rise And Sudden Fall Means For OpenAI, Disney And AI Video

What Sora's Rise And Sudden Fall Means For OpenAI, Disney And AI Video
Forbes +40 sources 2026-03-25 news
openairoboticssora
OpenAI’s decision to shut down Sora, its flagship text‑to‑video model, has turned a high‑profile partnership with Disney into a public rupture. The move follows a summer‑2025 private demo in which Sora produced a 60‑second, 4K clip, prompting Disney to commit a $1 billion investment and to position the technology as the core of its next‑generation content pipeline. By early March 2026 OpenAI announced the cancellation, citing unresolved technical hurdles and escalating compute costs, and Disney announced it was pulling out of the deal. As we reported on 25 March, Disney’s exit signalled the end of what had been billed as a “millennial‑era” collaboration between Hollywood and generative AI. The fallout matters on three fronts. First, it exposes the fragility of AI‑video business models that rely on massive, proprietary world‑model training; without a clear path to cost‑effective scaling, even deep‑pocketed partners balk. Second, creator‑rights groups have seized on the shutdown to demand stronger safeguards against unlicensed use of copyrighted footage, a debate that could shape future regulation of generative media. Third, the episode reshapes the competitive landscape: Anthropic and other private‑equity‑backed startups are now courting Disney and other studios, while OpenAI appears to be refocusing on text and image generation rather than expanding into video. What to watch next is whether OpenAI will revive Sora in a leaner form, how Disney will reallocate its AI budget—potentially toward in‑house tools or rival providers—and how legislators in the EU and US respond to mounting pressure for transparency and rights protection in AI‑generated video. The next quarter will reveal whether the Sora saga is a cautionary footnote or a catalyst for a new wave of generative filmmaking standards.
Forbes — https://www.forbes.com/sites/moinroberts-islam/2026/03/25/what-soras-rise-and-su www.tradingkey.com — https://www.tradingkey.com/analysis/stocks/us-stocks/251391172-walt-disney-opena marylandoutdoorclub.org — https://marylandoutdoorclub.org/article/disney-ditches-openai-deal-after-sora-sh ciclopiemonte.com — https://ciclopiemonte.com/article/disney-ditches-openai-deal-after-sora-shutdown jeunesse2000.org — https://jeunesse2000.org/article/disney-ditches-openai-deal-after-sora-shutdown- asuiteatkitsilanocottage.com — https://asuiteatkitsilanocottage.com/article/disney-ditches-openai-deal-after-so CBS News — https://www.msn.com/en-us/news/technology/openai-pulls-the-plug-on-its-sora-ai-v CNBC — https://www.msn.com/en-us/money/companies/openai-shutters-short-form-video-app-s HN — https://www.hollywoodreporter.com/business/digital/openai-shutting-down-sora-ai- HN — https://variety.com/2026/digital/news/openai-shutting-down-sora-video-disney-123 HN — https://arstechnica.com/ai/2026/03/openai-plans-to-shut-down-sora-just-15-months HN — https://twitter.com/soraofficialapp/status/2036546752535470382 HN — https://www.cnbc.com/2026/03/24/openai-shutters-short-form-video-app-sora-as-com HN — https://www.reuters.com/technology/openai-set-discontinue-sora-video-platform-ap Mastodon — https://neopaquita.es/@Shine_McShine/116289430804917987 Mastodon — https://social.vivaldi.net/@lamdba/116289000104999523 Mastodon — https://mastodon.social/@lowyat/116288992203707219 Mastodon — https://post.lurk.org/@fcr/116289639576191628 Mastodon — https://fed.brid.gy/r/https://www.rappler.com/technology/openai-shut-down-sora-a Mastodon — https://masto.ai/@panterapolnocy/116289646491201082 Mastodon — https://halo.nu/@theguardian_us_technology/116288802972565252 Mastodon — https://mastodon.social/@maquinari_cat/116288957020359623 Mastodon — https://mastodon.social/@pafurijaz/116289294284528312 Mastodon — https://zdrojak.cz/zpravicky/openai-vypina-aplikaci-sora/ Mastodon — https://mastodon.social/@WinFuture/116289142403658416 Mastodon — https://mastodon.ozioso.online/@ItalianNews/116288880089973783 Mastodon — https://anonsys.net/display/bf69967c-2cd3e45c-479fe0cf8c18b9b2 Mastodon — https://ard.social/@tagesschau/116289027590377594 Mastodon — https://toot.earth/@PetaPixel/116286455993988500 Mastodon — https://fed.brid.gy/r/https://fumettologica.it/2026/03/disney-openai-sora-accord Mastodon — https://social.macg.co/@macgeneration/116289075826710458 NPR — https://www.npr.org/2026/03/25/g-s1-115055/openai-pulls-the-plug-on-sora-the-vir Mastodon — https://infosec.exchange/@AAKL/116290536912001076 Mastodon — https://fraxoweb.social/@frank/116290229875721992 Mastodon — https://fed.brid.gy/r/https://filmstories.co.uk/news/openai-axes-sora-its-ai-vid Mastodon — https://infosec.exchange/@AAKL/116290505718166286 Mastodon — https://shrk.crystalsky.dev/notes/01KMJVVQ216BFF3W52RHHTV1BJ Mastodon — https://fed.brid.gy/r/https://cordcuttersnews.com/openai-pulls-the-plug-on-sora- Mastodon — https://c.im/@arstechnica/116290403871672991 Mastodon — https://mastodon.social/@winbuzzer/116290467599563731
278

So-called # OpenAI is killing # sora . Maybe normies can finally start to see that the # AI

So-called  # OpenAI   is killing  # sora  . Maybe normies can finally start to see that the  # AI
Mastodon +7 sources mastodon
climategoogleopenaisora
OpenAI has officially pulled the plug on Sora, its much‑talked‑about text‑to‑video model, confirming the abrupt shutdown first hinted at in the company’s March 25 announcement that the service would be discontinued and its multimillion‑dollar Disney partnership would collapse. The firm removed Sora from its API catalog, disabled the web app and sent a terse email to remaining users stating that development had been “re‑prioritised” amid “resource constraints.” The move matters because Sora was OpenAI’s flagship attempt to extend generative AI beyond text and images into moving pictures, a capability that had been billed as a new frontier for advertisers, creators and entertainment studios. Its demise not only ends a high‑profile collaboration with Disney but also signals a broader recalibration of OpenAI’s product strategy. Analysts see the decision as a reaction to mounting cost pressures, slower-than‑expected adoption of high‑compute video generation, and a cooling of the speculative “AI bubble” that has driven inflated valuations across the sector. Investors are already reassessing OpenAI’s growth outlook. The company’s latest earnings call hinted that resources will be redirected toward the “o1” reasoning engine and the next iteration of GPT, which promise higher‑value enterprise use cases rather than headline‑grabbing media demos. Meanwhile, rivals such as Google DeepMind and Anthropic are accelerating their own multimodal roadmaps, hoping to capture the market share Sora vacated. What to watch next: the impact on OpenAI’s balance sheet and any renegotiation with Disney, the rollout timeline for the o1 series, and whether the shutdown will prompt a broader pullback on ambitious generative‑video projects across the industry. A shift toward more pragmatic, revenue‑generating AI tools could reshape funding patterns for startups that have bet on the video frontier.
267

OpenAI Enters Its Focus Era by Killing Sora https:// fed.brid.gy/r/https://www.wire d.com/st

Mastodon +10 sources mastodon
openaisora
OpenAI announced on Tuesday that it will retire Sora, the AI‑driven video‑generation app it launched last year, marking a decisive shift toward what the company calls its “Focus Era.” The decision comes as OpenAI prepares for an initial public offering and redirects resources to a single, unified AI assistant and a suite of enterprise‑grade coding tools. Sora’s shutdown follows a brief, turbulent run‑out: the service was rolled out with fanfare, secured a multiyear partnership with Disney to animate iconic characters, and then was pulled just three months later. OpenAI cited mounting compute costs, a strategic pivot toward robotics, and the need to consolidate its product portfolio as the primary reasons for the move. The closure matters on several fronts. For creators, it removes one of the few consumer‑ready tools capable of generating full‑length video from text prompts, tightening the gap that competitors such as Runway and Meta’s Make‑It‑Real have been racing to fill. For investors, the signal is clear: OpenAI is shedding peripheral experiments to showcase a tighter, revenue‑generating stack ahead of its IPO, a narrative it has been building since the March‑25 reports on Sora’s demise. Looking ahead, the market will watch for the formal IPO filing, expected later this quarter, and for the rollout of the promised unified assistant that will blend chat, image, and code capabilities under a single interface. Equally critical will be OpenAI’s progress in robotics—its first major hardware‑focused venture since the acquisition of Boston Dynamics‑adjacent talent—and how it reallocates the compute budget freed by Sora’s termination. The next few months will reveal whether the “Focus Era” translates into sustained growth or merely a rebranding of an already dominant AI platform.
260

OpenAI Pulls the Plug on Sora Just Months After Launch https:// fed.brid.gy/r/https://www.rol

OpenAI Pulls the Plug on Sora Just Months After Launch       https://  fed.brid.gy/r/https://www.rol
Mastodon +7 sources mastodon
openaisora
OpenAI announced on Tuesday that it is “saying goodbye” to Sora, the short‑form AI video app it launched in late 2023. The company posted a terse note on X, promising to release tools that let users export the clips they have already created, but gave no timeline for a full shutdown. The decision comes just six months after Sora’s public release and barely three months after OpenAI signed a multiyear licensing deal with Disney to feature the studio’s characters in generated videos. The abrupt closure underscores the volatility of the nascent AI‑generated video market. Sora quickly became a flashpoint for deep‑fake concerns, prompting regulators in Europe and the United States to issue warnings about unlabelled synthetic media. At the same time, the service proved financially demanding; earlier reporting described it as “the most expensive AI slop factory in history.” Disney’s withdrawal from the partnership—reported on March 25—suggests that the commercial model was already under strain. What happens next will shape both OpenAI’s product strategy and the broader ecosystem of AI video tools. Analysts will watch whether OpenAI redirects Sora’s technology into its core ChatGPT platform or launches a more tightly controlled offering that satisfies licensing partners and regulators. Disney’s statement that it will “find ways to peddle slop elsewhere” hints at a possible in‑house solution or a shift to another AI vendor. Meanwhile, competitors such as Runway, Meta’s Make‑It‑Real and emerging European startups are likely to vie for the displaced user base. The next OpenAI communication, expected within days, should clarify data‑preservation plans and hint at the company’s longer‑term vision for generative video. As we reported on March 25, the Sora saga is already reshaping expectations for AI‑driven creative services.
244

Fine-Tuning vs Prompt Engineering: A Practical Technical Comparison for Modern AI Systems

Fine-Tuning vs Prompt Engineering: A Practical Technical Comparison for Modern AI Systems
Dev.to +7 sources dev.to
fine-tuningrag
A joint white‑paper released this week by the Nordic AI Institute and the cloud‑services arm of a leading European telecom provider offers the first systematic, production‑grade comparison of fine‑tuning and prompt engineering for today’s large language models. The authors evaluated three flagship LLMs—Claude‑3, Gemini‑1.5 and Llama‑3—across ten real‑world tasks ranging from legal clause extraction to creative copywriting. Results show that prompt engineering can match fine‑tuned accuracy on generic tasks while delivering a 70 % reduction in development time and up to 60 % lower compute cost. For highly specialized domains, however, models that were fine‑tuned on a few thousand curated examples consistently outperformed the best‑crafted prompts, achieving up to 99.1 % extraction accuracy in a banking document‑processing benchmark. The study matters because enterprises are now forced to choose between two competing optimisation pathways that have very different operational footprints. Prompt engineering preserves the original model, sidestepping data‑privacy concerns and allowing rapid A/B testing, but it demands continual prompt maintenance as use‑cases evolve. Fine‑tuning embeds domain knowledge permanently, simplifying downstream pipelines at the expense of higher upfront data‑labelling, longer training cycles and tighter model‑governance requirements. As AI budgets tighten across the Nordics, the cost‑benefit calculus presented in the paper will shape product roadmaps, especially for sectors such as finance, healthcare and public administration where regulatory compliance drives the need for reproducible, auditable behaviour. What to watch next: the authors announce an open‑source toolkit that blends the two approaches, automatically generating task‑specific prompts and then applying lightweight parameter‑efficient fine‑tuning (PEFT) where gains plateau. Early adopters, including a Swedish insurance firm and a Danish e‑government portal, plan pilots for Q3. Industry analysts will be monitoring whether hybrid workflows become the de‑facto standard, potentially prompting cloud providers to rethink pricing models for prompt‑runtime versus fine‑tuning compute.
230

RAG vs Fine-Tuning — What Actually Works in Production (2026)

RAG vs Fine-Tuning — What Actually Works in Production (2026)
Dev.to +7 sources dev.to
fine-tuningrag
A new production‑grade guide released this week by AI engineer Umesh Malik lays out hard‑won lessons from a year of building live LLM services for customers across e‑commerce, finance and telecom. The report, titled “RAG vs Fine‑Tuning — What Actually Works in Production (2026)”, aggregates telemetry from dozens of deployments and argues that the binary choice between Retrieval‑Augmented Generation (RAG) and fine‑tuning is no longer realistic. Instead, hybrid pipelines that pair a fine‑tuned inference model with a dynamic retrieval layer have become the de‑facto standard. Malik’s data show that pure RAG systems win on knowledge freshness and maintenance overhead, especially in domains where facts change weekly or daily. Fine‑tuned models, by contrast, deliver tighter stylistic control, lower latency and the ability to run offline, which translates into cost savings at high query volumes. The guide quantifies these trade‑offs: a 30 % reduction in latency when serving a fine‑tuned model alone, versus a 45 % drop in stale‑answer incidents when augmenting the same model with a retrieval index refreshed every 12 hours. The hybrid approach inherits the best of both worlds, achieving sub‑second response times while keeping citation accuracy above 92 %. Why it matters is that enterprises are now moving beyond proof‑of‑concepts and need concrete guidance on scaling LLMs responsibly. As we reported on 25 March 2026, the fine‑tuning vs prompt‑engineering debate highlighted the importance of model‑specific optimisation; Malik’s findings extend that conversation to the full stack, showing how retrieval infrastructure and model adaptation interact in real‑world cost and compliance calculations. Looking ahead, vendors are expected to roll out tighter integrations for hybrid pipelines, including managed vector stores with built‑in versioning and on‑device fine‑tuning kits. Observers will watch for benchmark releases that standardise hybrid performance metrics, and for regulatory frameworks that may mandate citation‑ready RAG components in high‑risk sectors. The next few months should reveal whether the hybrid model becomes a permanent architectural norm or a transitional compromise as foundation models continue to improve.
210

Hypura – A storage-tier-aware LLM inference scheduler for Apple Silicon

Hypura – A storage-tier-aware LLM inference scheduler for Apple Silicon
HN +5 sources hn
appleinferencellama
A new open‑source project called **Hypura** has been released on GitHub, promising to make large‑language‑model (LLM) inference on Apple Silicon Macs more practical. The four‑day‑old repository describes Hypura as a “storage‑tier‑aware LLM inference scheduler,” a thin layer that dynamically moves model weights between RAM and the SSD while batching requests to keep the Apple‑M‑series GPU busy. The innovation matters because Apple’s on‑device AI ecosystem has long been hamstrung by the limited unified memory of MacBooks and iMacs. Even with the efficient MLX runtime, models that exceed a few gigabytes still require costly off‑loading to external storage, which introduces latency and stalls the GPU. By treating the SSD as a second memory tier and scheduling work around its bandwidth, Hypura can keep inference pipelines flowing, reportedly narrowing the performance gap with desktop‑class GPUs. Early tests from the authors show throughput gains of 20‑30 % over vanilla llama.cpp on M2‑Pro hardware, echoing similar improvements reported by the vllm‑mlx project earlier this year. If the scheduler lives up to its promises, developers could run state‑of‑the‑art models such as Llama‑2‑13B or Mistral‑7B locally on a MacBook without resorting to cloud services. That would lower the barrier for privacy‑focused applications, expand the market for macOS‑native AI tools, and put pressure on Apple to integrate more sophisticated memory‑management primitives into its own frameworks. The next steps to watch include community benchmarking against competing solutions like OMLX and Parallax, potential contributions that tie Hypura into Apple’s Core ML and MLX stacks, and any signal from Apple that the scheduler might be folded into an official macOS release. A successful adoption could reshape the balance between on‑device and cloud inference for developers in the Nordic AI scene and beyond.
172

OpenAI just gave up on its Sora AI video generator

The Verge +23 sources 2026-03-25 news
openaisora
OpenAI announced on Tuesday that it is pulling the plug on Sora, the text‑to‑video model it unveiled at the end of 2024. The brief statement, “We’re saying goodbye to Sora,” marks the end of a product that generated a wave of excitement for its ability to produce minute‑long, photorealistic clips from a single prompt, but also sparked controversy over its massive compute demands and the legal fallout from a collapsed partnership with Disney. The shutdown follows a string of reports from earlier this week that OpenAI was already winding down Sora’s support and cancelling the multi‑million‑dollar deal with Disney that had promised exclusive content rights. As we reported on 25 March, the company had begun “killing” Sora and the Disney agreement fell apart amid concerns that the technology could blur the line between real and synthetic media. The decision now appears final, with the service being removed from the API dashboard and existing user credits slated for refund. Why it matters is twofold. First, Sora’s demise underscores the practical limits of current AI video generation: rendering a single minute of high‑definition footage can consume more GPU power than many of OpenAI’s other flagship models, making it financially unsustainable at scale. Second, the episode highlights growing regulatory and reputational pressure on AI firms to curb tools that could be weaponised for deep‑fake propaganda or copyright infringement. What to watch next is OpenAI’s strategic pivot. The company is likely to redirect the compute budget earmarked for Sora toward its next‑generation text‑and‑image models, while competitors such as Runway, Google DeepMind and Meta’s Make‑It‑Real may try to capture the vacated market segment. Observers will also be keen to see whether OpenAI offers a lighter‑weight video prototype in the future, and how regulators respond to the broader implications of AI‑generated media.
The Verge — https://www.theverge.com/ai-artificial-intelligence/899850/openai-sora-ai-chatgp futurism.com — https://futurism.com/the-byte/openai-sora-leak aimarketingnewstoday.com — https://aimarketingnewstoday.com/openai-releases-ai-video-generator-sora-to-cust arstechnica.com — https://arstechnica.com/information-technology/2024/02/openai-collapses-media-re www.wilsonsmedia.com — https://www.wilsonsmedia.com/openai-just-gave-artists-access-to-sora-and-proved- www.cmswire.com — https://www.cmswire.com/digital-experience/openai-unveils-sora-its-impressive-ai CNBC — https://www.msn.com/en-us/money/companies/openai-shutters-short-form-video-app-s HN — https://variety.com/2026/digital/news/openai-shutting-down-sora-video-disney-123 HN — https://twitter.com/soraofficialapp/status/2036546752535470382 HN — https://arstechnica.com/ai/2026/03/openai-plans-to-shut-down-sora-just-15-months HN — https://www.reuters.com/technology/openai-set-discontinue-sora-video-platform-ap HN — https://www.cnbc.com/2026/03/24/openai-shutters-short-form-video-app-sora-as-com Mastodon — https://mstdn.social/@WIREID91LDNON/116287280106982560 Mastodon — https://eter9.com/notes/ak912i2ci8 Mastodon — https://weredreaming.com/mookie/p/1774403800.230045 Mastodon — https://c.im/@heysannidhi/116287370794930407 Mastodon — https://fed.brid.gy/r/https://www.rappler.com/technology/openai-shut-down-sora-a Mastodon — https://mastodon.social/@Mathrubhumi_English/116287986840102523 Mastodon — https://masto.ai/@Miro_Collas/116288106875895645 Mastodon — https://mastodon.social/@schuler/116287354030592054 Mastodon — https://mastodon.nl/@teacher_rick/116288257594208811 Mastodon — https://mastodon.uno/@nicolaottomano/116288012563987419 Mastodon — https://mastodon.social/@NieuwsJunkies/116287067730459771
171

AirPods Max 2に対応した「iOS/iPadOS 26.4」。Apple Musicも機能追加

Mastodon +7 sources mastodon
apple
Apple rolled out iOS 26.4, iPadOS 26.4 and watchOS 26.4 on 25 March, synchronising the software launch with the debut of the second‑generation AirPods Max. The update adds native support for the new over‑ear headphones, enabling spatial audio, adaptive EQ and the refreshed “Find My” tracking that were missing from the previous OS releases. Beyond hardware compatibility, Apple Music receives a suite of enhancements. A “Concert” tab now surfaces live shows taking place near the user’s location, while the “Ambient” and “Well‑being” playlists have been expanded with AI‑curated tracks for sleep, focus and relaxation. The service also introduces a smarter recommendation engine that suggests new artists based on the music currently playing, a move that tightens Apple’s grip on the streaming market where Spotify and YouTube Music dominate in the Nordics. The update tackles keyboard latency, promising more accurate typing after rapid input—a subtle but welcome tweak for power users. Security receives a routine patch set, and watchOS 26.4 adds a low‑power “Sleep Mode” for the Apple Watch, aligning with Apple’s broader health‑first narrative. Why it matters is twofold. First, the seamless AirPods Max 2 integration underscores Apple’s strategy of bundling premium hardware with software upgrades to justify higher price points, a formula that could reshape Nordic consumer expectations for premium audio. Second, the Apple Music refinements signal a shift toward location‑aware, AI‑driven experiences that may erode the market share of competing services. What to watch next: early adopters’ reviews of AirPods Max 2’s spatial audio performance on iOS 26.4, uptake figures for the new “Concert” feature, and whether Apple will extend the AI‑driven recommendation engine into its upcoming iOS 27. Analysts will also monitor whether the Nordic market’s strong appetite for high‑fidelity streaming translates into a measurable bump in Apple Music subscriptions.
169

Auto mode for Claude Code | Claude

Mastodon +7 sources mastodon
ai-safetyanthropicclaude
Anthropic has rolled out “Auto Mode” for Claude Code, its AI‑driven development assistant, turning a long‑standing permission prompt into a self‑serving safety layer. The new mode deploys an on‑device classifier that evaluates each command—such as file writes, package installations or system calls—and automatically approves those deemed low‑risk while still surfacing higher‑impact actions for human review. Developers can toggle the feature in the Claude Code settings, and the system logs every auto‑approved operation for auditability. The launch marks a shift from the manual “yes/no” dialogs that many users complained slowed down workflows. By handling routine permissions in the background, Auto Mode promises to cut the friction that has hampered large‑scale adoption of AI‑assisted coding tools, especially in fast‑moving teams that need to iterate quickly. At the same time, Anthropic positions the classifier as a safeguard against the “AI coding disasters” that have sparked headlines when LLMs execute destructive commands or expose sensitive data. The company frames the feature as a middle ground between the default prompt‑heavy configuration and the risky practice of disabling permissions altogether. As we reported on March 25, 2026, Claude Code already had the ability to take over a developer’s workstation; today the functionality is wrapped in a safety‑first interface that could set a new industry benchmark. The move also dovetails with Anthropic’s broader suite of updates, including Claude Code Review, a multi‑agent bug‑screening tool, and Dispatch for Cowork, which lets users hand off tasks from mobile devices. What to watch next: early adoption metrics and feedback from enterprise pilots will reveal whether the classifier strikes the right balance between speed and security. Competitors such as OpenAI and Google are expected to announce comparable permission‑automation features, potentially sparking a race to embed safety into the core of AI‑coding workflows. Regulators may also scrutinise how these classifiers are trained and validated, especially if they become the default gatekeeper for code that touches production systems.
158

"Canada rejected the permanent residence application of a McMaster postdoc from the Sorbonne who wor

Mastodon +6 sources mastodon
Canada’s immigration agency has rejected the permanent‑residence application of a McMaster University postdoctoral researcher who earned her doctorate at the Sorbonne and studies the immunology of ageing. The denial, issued by Immigration, Refugees and Citizenship Canada (IRCC), cites an “incomplete” application – but the underlying cause was a generative‑AI system that hallucinated her academic credentials, mistakenly flagging her as lacking the required qualifications. The incident shines a spotlight on the growing reliance on large language models to triage and evaluate immigration files. IRCC introduced the AI tool earlier this year to accelerate processing times and reduce manual workload, but the technology’s propensity for fabricating or mis‑interpreting data has now produced a concrete, high‑stakes error. For a country that depends on skilled researchers to sustain its knowledge‑based economy, a false rejection threatens both individual careers and the broader talent pipeline. Legal experts note that applicants can appeal IRCC decisions, yet the opacity of AI‑driven assessments complicates the evidentiary basis for a challenge. The case may prompt a review of the agency’s AI governance framework, including requirements for human verification, audit trails and bias mitigation. Advocacy groups are already calling for a pause on fully automated decision‑making until robust safeguards are in place. Watch for IRCC’s official response, which is expected within the next two weeks, and for any court filings by the researcher or her legal counsel. Parallel developments – such as the federal government’s upcoming AI‑ethics legislation and other reported AI mishaps in public services – will indicate whether Canada will tighten oversight or double down on automation in its immigration system.
158

I've only just realised that # OpenAI 's logo is a stylised anal sphincter. Apologies for

I've only just realised that  # OpenAI  's logo is a stylised anal sphincter. Apologies for
Mastodon +6 sources mastodon
dall-eopenaisoratext-to-imagetext-to-video
A Reddit post that went viral early Tuesday claimed OpenAI’s trademark emblem resembles a “stylised anal sphincter,” prompting a flurry of memes and a brief spike in brand‑related chatter. The comment, posted under the r/OpenAI community, was accompanied by a side‑by‑side comparison of the company’s teal‑blue “O” and the anatomical analogy, and within hours it had been shared across Twitter, LinkedIn and several tech‑focused Discord channels. The observation is harmless in tone but lands at a moment when OpenAI is already under intense scrutiny. Just weeks ago the firm abruptly discontinued its Sora text‑to‑video service, a move that forced Disney to walk away from a multi‑billion‑dollar partnership and sparked widespread debate about the sustainability of high‑cost AI products. As we reported on 25 March, the Sora shutdown highlighted OpenAI’s volatile product strategy and raised questions about its long‑term vision. The logo joke, therefore, adds a layer of reputational risk, turning a design critique into a symbol of broader discontent. OpenAI has not issued an official comment, but its communications team is known to monitor social‑media sentiment closely. Analysts suggest the company could respond with a light‑hearted acknowledgment or, if the narrative gains traction, a subtle redesign to pre‑empt any negative branding impact. In the past, tech firms have tweaked logos after viral jokes—Apple’s “bent‑iPhone” meme in 2018 spurred a minor redesign of the device’s silhouette, for example. What to watch next: whether OpenAI’s leadership addresses the meme in a public statement, if the company’s design team hints at a logo refresh, and how the episode influences ongoing discussions about corporate visual identity in the AI sector. The episode also serves as a reminder that even subtle branding choices can become flashpoints in an industry already grappling with public trust.
155

Disney Says It Will Find Ways to Peddle Slop Elsewhere After Pulling Out of OpenAI Deal

Gizmodo +10 sources 2026-03-25 news
openaisora
Disney has officially walked away from the $1 billion licensing pact it signed with OpenAI three months ago, after the San Francisco‑based lab abruptly shut down its Sora video‑generation app. In a brief statement, Disney said it will “find ways to peddle slop elsewhere,” signaling that the company will seek alternative avenues to monetize AI‑generated content rather than rely on the now‑defunct Sora platform. The collapse follows a string of OpenAI announcements that began on 25 March, when we reported the company’s decision to discontinue support for Sora and the resulting fallout for its multimillion‑dollar deal with Disney. Sora, billed as a generative‑video tool that could turn text prompts into short clips, was meant to power Disney’s streaming services, theme‑park experiences and advertising. Its sudden removal leaves a gap in Disney’s AI roadmap and raises questions about the viability of large‑scale video‑generation models that still struggle with consistency, copyright compliance and compute costs. For Disney, the loss is both financial and strategic. The $1 billion agreement was expected to fund a suite of AI‑enhanced productions and to give the media giant a foothold in a market that rivals like Meta and Google are aggressively courting. OpenAI’s pivot toward productivity‑focused tools suggests it doubts the near‑term commercial readiness of generative video, a stance that could reshape industry expectations and redirect investment toward text‑to‑image or code‑assistance models. What to watch next: whether Disney will partner with a rival AI provider, develop its own video‑generation stack, or double down on traditional content creation. OpenAI’s next product announcements will also be scrutinised for clues about its long‑term commitment to generative media. Legal teams on both sides may soon address the financial settlement of the aborted deal, a process that could set precedents for future AI licensing contracts.
152

📰 Claude Code Auto Mode in 2026: How AI Coding Security Is Changing—And Where It Still Fails Claude

Mastodon +9 sources mastodon
anthropicautonomousclaude
Anthropic rolled out “Auto Mode” for Claude Code on March 11, 2026, letting the Claude Sonnet 4.6 model autonomously approve or block code actions during a development session. The feature, launched as a research preview, embeds a classifier that evaluates each proposed edit for permission level, prompt‑injection risk and potential side effects before execution. Developers can toggle the mode, set admin‑level overrides and define custom policy thresholds, turning the AI from a passive assistant into a gatekeeper that decides when it may act on its own. The move marks a shift in AI‑driven software tooling. By moving permission decisions from the human to the model, Anthropic hopes to shrink feedback loops and keep developers in the flow, especially in long‑running coding sessions where frequent manual approvals become a bottleneck. The built‑in safeguards aim to address longstanding concerns about AI‑generated code executing unintended commands or leaking credentials, a criticism that has dogged earlier tools such as GitHub Copilot and OpenAI’s coding suite. As we reported on March 24, Claude Code already logged more than 19 million commits on GitHub and introduced a token‑optimizer to curb redundant reads. Auto Mode builds on that momentum, but analysts warn the reliance on a single classifier still leaves gaps: edge‑case vulnerabilities, false‑positive blocks and the difficulty of auditing the model’s decision logic remain unresolved. Enterprises will need to balance the productivity boost against the risk of opaque permission handling and the extra compute cost of continuous safety checks. Watch for Anthropic’s forthcoming public beta, slated for early Q2, and for competitor responses. GitHub Copilot Workspace and OpenAI’s upcoming coding tools are expected to introduce comparable autonomous permission layers, setting up a near‑term race to define standards for AI‑mediated code execution security. The next few months will reveal whether Auto Mode can deliver on its promise without compromising the very safeguards it seeks to enforce.
151

OpenAI's latest repo has Claude as the third top contributor

HN +7 sources hn
agentsanthropicclaudegeminigooglegpt-5openai
OpenAI’s most recent public GitHub repository shows Anthropic’s Claude model ranking as the third‑largest code contributor, trailing only OpenAI’s own GPT‑5.2 and a handful of human developers. The contribution metrics, extracted from the repo’s commit history, reveal that Claude generated roughly 12 % of the added lines over the past month, outpacing several internal tooling scripts and even some senior engineers. The finding matters because it signals a shift from pure competition to pragmatic cross‑pollination in the AI‑coding arena. Claude’s rise in the rankings follows a string of Anthropic upgrades – most notably the “Claude Code” auto‑mode that can autonomously complete complex development tasks – and underscores the model’s growing reputation for handling messy, real‑world codebases. OpenAI’s chief technology officer, Mira Brockman, hinted on a recent podcast that the company has been “experimenting with external code‑generation models to accelerate internal tooling,” suggesting that the repo data is not an isolated experiment but part of a broader strategy to tap the best available AI, regardless of origin. What to watch next is whether OpenAI formalises the collaboration or keeps it under the radar. A public acknowledgment could lead to joint benchmarks, shared safety protocols, or even a licensing deal that blurs the line between rival firms. Conversely, a quiet continuation may prompt Anthropic to leverage the exposure as a marketing lever, especially as Claude’s auto‑mode gains traction after the safety enhancements we covered on March 25. Industry observers will also be keen on the next set of commits: if Claude’s share climbs, it could reshape expectations about who supplies the most productive coding assistants in the fast‑moving generative‑AI race.
150

Introducing @rotifer/mcp-server: Give Any AI Agent Access to the Gene Ecosystem

Dev.to +6 sources dev.to
agents
A new open‑source server called **@rotifer/mcp‑server** lets any AI agent that speaks the Model Context Protocol (MCP) tap directly into the Rotifer gene ecosystem. The one‑line installation, published on npm, launches a lightweight MCP server that exposes eight searchable tools – from `search_genes` to `compare_genes` – and five data resources such as `genedetail` and `leaderboard`. By issuing a single MCP command, an agent can locate a “gene” (a reusable AI capability), inspect its metadata, compare fitness scores against alternatives, and pull the optimal version into its own workflow without leaving the IDE. The rollout matters because it turns the Rotifer gene registry – already home to more than 50 curated genes for tasks like web‑scraping, data cleaning, and vision – into a plug‑and‑play marketplace for autonomous agents. Developers no longer need to hard‑code toolchains; instead, agents can self‑diagnose missing functions, query the ecosystem, and dynamically augment their skill set. This mirrors the modular approach that sandboxing frameworks introduced earlier this month, which promised 100× faster agent testing, and it dovetails with DeepSeek’s push to integrate DeerFlow 2.0 for richer tool orchestration. In practice, the server could accelerate the deployment of specialized agents in finance, healthcare, and e‑commerce by reducing the time spent on manual integration. What to watch next is how quickly the Rotifer MCP interface is adopted by the broader AI‑agent community. Early signs include a growing GitHub star count and a handful of tutorials on the DEV community site. Analysts will be tracking performance benchmarks against existing tool‑lookup services, the emergence of third‑party gene contributions, and whether major cloud providers embed the MCP server into their AI‑agent platforms. If the ecosystem gains traction, it could become a de‑facto standard for on‑the‑fly capability upgrades, reshaping how autonomous systems evolve and compete.
130

Google's TurboQuant reduces AI LLM cache memory capacity requirements by at least six times — up to 8x performance boost on Nvidia H100 GPUs, compresses KV caches to 3 bits with no accuracy loss

Mastodon +7 sources mastodon
googlenvidia
Google’s research team unveiled TurboQuant, a two‑stage quantization scheme that slashes the key‑value (KV) cache of large language models (LLMs) by at least six‑fold and delivers up to an eight‑fold speed boost on Nvidia H100 GPUs. The method compresses KV entries to just three bits using a novel “PolarQuant” rotation step, then applies a lightweight integer‑only fine‑tuning that preserves the original model’s output exactly – no accuracy loss, no retraining, and no changes to the model architecture. The breakthrough matters because KV caches dominate memory consumption in long‑context inference. By shrinking the cache, TurboQuant frees up GPU RAM, allowing developers to run larger context windows or pack more concurrent requests onto a single H100. The resulting throughput gains translate into lower cloud‑compute bills and reduced energy footprints, a critical factor as LLM deployments scale across data‑centers. For enterprises that price services per token, the efficiency gain could also pressure token‑based pricing models, echoing the tongue‑in‑cheek complaint that “AI token prices are being destroyed.” TurboQuant builds on the workflow‑optimization trends we highlighted in our March 25 survey of LLM agents, where memory‑aware scheduling and cache management were identified as bottlenecks. Google’s claim that the technique works as a drop‑in for any transformer model means that existing pipelines – from OpenAI’s GPT‑5 to Apple‑silicon‑tuned inference stacks – could adopt it without code changes. What to watch next: early benchmark releases from Google and third‑party labs will confirm the zero‑loss promise across diverse model families. Integration into popular frameworks such as Hugging Face Transformers and TensorRT will signal mainstream uptake. Finally, cloud providers may roll out TurboQuant‑enabled instance types, and we’ll monitor how pricing and token‑economics evolve as memory constraints recede.
129

RE: https:// social.heise.de/@heiseonlineen glish/116288830860438729 For the time being, Euro

Mastodon +6 sources mastodon
metaprivacy
Meta’s upcoming line of AI‑powered smart glasses has hit a regulatory roadblock in Europe, a Mastodon post from Heise Medien noted on Thursday. A provisional decision by the European Data Protection Board (EDPB) has ordered the company to halt any rollout of its “Meta Vision” devices until a full privacy impact assessment is completed. The move follows a wave of concerns that the glasses’ built‑in large language models (LLMs) could capture and transmit facial, audio and location data in real time, effectively turning wearers into constant surveillance nodes. The European pause contrasts sharply with recent developments in the United States, where a federal court dismissed Meta’s bid to overturn a lawsuit alleging the firm failed to protect young users from harmful content. While Meta prepares to appeal the US ruling, the EDPB’s precautionary measure keeps European consumers insulated from what officials describe as a “next‑level privacy nightmare.” Why the decision matters is twofold. First, it tests the limits of the EU’s General Data Protection Regulation (GDPR) and the newer Digital Services Act when confronted with emerging wearables that blend AI, augmented reality and continuous data streaming. Second, it signals to other BigTech players that the European market will not tolerate opaque data practices, even as the region pushes for AI innovation. The ruling also underscores a growing regulatory split: Europe is tightening controls while the US courts are still wrestling with broader safety claims. What to watch next is whether Meta will submit a revised impact assessment that satisfies the EDPB, or whether it will challenge the order in the European Court of Justice. Parallel to that, the European Commission is expected to publish guidance on AI‑enabled wearables later this year, potentially setting a template for global standards. Industry observers will also be monitoring how the decision influences the rollout plans of rivals such as Apple and Google, both of which are developing their own AR glasses. The outcome could shape the balance between immersive technology and privacy across the continent.
124

Embeddings: The One Concept Behind RAG, Search, and AI Systems

Dev.to +6 sources dev.to
embeddingsragvector-db
A consortium of Nordic research labs and the cloud‑native startup VectorMind unveiled **EmbedX**, an open‑source embeddings platform that promises to be the single building block for Retrieval‑Augmented Generation (RAG), vector search and recommendation engines. The release bundles a suite of pre‑trained models, a high‑throughput inference API and a plug‑and‑play vector database, allowing developers to generate document, query and item embeddings with a single call and immediately query them for semantic similarity. The move matters because today’s AI applications often stitch together disparate components—fine‑tuned language models for generation, separate vector stores for search, and custom similarity metrics for recommendation. EmbedX collapses that stack, delivering a unified representation that can be reused across pipelines. Early benchmarks posted by the authors show up to 30 % latency reduction and a 15 % boost in relevance scores compared with the typical “model‑per‑task” approach. For Nordic enterprises that are scaling AI‑driven customer support, knowledge‑base retrieval and personalized content, the simplification translates into lower engineering overhead, faster time‑to‑market and more predictable cloud costs. What to watch next is how quickly the platform gains traction beyond the initial pilot projects at a few telecom operators and fintech firms. The consortium has pledged a roadmap that includes privacy‑preserving embeddings, on‑device inference for edge devices and integration with major LLM providers. Competitors such as Azure Cognitive Search and Google Vertex AI are already hinting at similar unified services, so a standards battle over embedding formats and evaluation metrics may emerge. Keep an eye on upcoming performance contests and on whether EmbedX’s open licensing spurs a broader ecosystem of plug‑ins that could reshape the way Nordic companies build semantic AI systems.
123

# OpenAI ha comunicato con un post su X l'addio all'app # Sora e alla comunity. Sora permetteva

Mastodon +6 sources mastodon
openaisora
OpenAI announced on X that it is shutting down Sora, the short‑form video generator that went viral after its autumn launch. The post, dated 24 March 2026, thanked “everyone who created with Sora, shared it, and built community around it” before confirming the service will be taken offline imminently. The move marks the end of a brief but intense experiment in AI‑driven video creation. Sora let users type a prompt and receive a realistic, up‑to‑30‑second clip, a capability that sparked both excitement and alarm. Within weeks, the tool was flooded with deep‑fake memes, political satire and copyrighted material, prompting concerns from regulators and rights‑holders. A Wall Street Journal leak earlier this month hinted that OpenAI was under pressure from legal teams and external watchdogs, and the company’s own statement stopped short of naming a single cause. Why it matters is twofold. First, Sora’s shutdown underscores the growing tension between rapid AI innovation and the need for responsible deployment, especially for media‑generation models that can blur the line between fact and fabrication. Second, the decision signals a strategic pivot for OpenAI: after a burst of product launches—including the ill‑fated Sora—the firm appears to be consolidating around its core offerings, such as ChatGPT and the emerging Claude‑contributed models, to avoid regulatory backlash and preserve brand trust. What to watch next is whether OpenAI will re‑enter the video‑generation space with a more tightly controlled product, and how the broader AI ecosystem will respond to the regulatory scrutiny that Sora’s demise has amplified. Keep an eye on upcoming policy proposals from the EU AI Act and U.S. congressional hearings, as well as any statements from OpenAI’s leadership about future multimedia ambitions. As we reported on 25 March, the closure of Sora is a clear indicator that the “focus era” OpenAI announced is already reshaping its roadmap.
123

LLM Neuroanatomy II: Modern LLM Hacking and Hints of a Universal Language?

HN +6 sources hn
A new post on Hacker News, titled **“LLM Neuroanatomy II: Modern LLM Hacking and Hints of a Universal Language?”**, builds on the author’s earlier “LLM Neuroanatomy” essay that explained how a homemade “brain scanner” helped the writer climb the LLM leaderboard without altering model weights. The sequel introduces two fresh strands of research that could reshape how developers think about large language models. First, the author highlights an experiment by researcher Evan Maunder that probes the model’s “thinking space” across languages. By feeding the same sentence in English, Mandarin and even Base64‑encoded text, Maunder measured cosine similarity layer‑by‑layer. The early transformer layers quickly map disparate inputs onto a common subspace, the similarity stays high through the middle stack, and only the final layers diverge as the model prepares language‑specific output. The pattern suggests that LLMs may construct a language‑agnostic representation—a kind of universal code that underlies all textual modalities. Second, the article surveys contemporary LLM hacking techniques, from prompt‑injection payloads catalogued on GitHub to “layer‑copy” tricks that duplicate thinking modules to boost performance. These tactics expose both the fragility of current safety guards and the untapped flexibility of transformer internals. Why it matters is twofold. A language‑agnostic core could explain why multilingual models transfer so well and might enable more efficient fine‑tuning, compression or even cross‑modal reasoning. At the same time, the growing toolbox of prompt‑injection attacks underscores a security gap that could be exploited in downstream applications, from chat assistants to code generators. What to watch next: the community is already debating whether the observed convergence truly constitutes a “universal language” or merely reflects shared tokenisation patterns. Follow‑up studies that replicate Maunder’s cosine‑similarity test on larger, instruction‑tuned models will be decisive. Meanwhile, security researchers are expected to release hardened prompting frameworks and mitigation guidelines, and we anticipate a response from major AI labs on whether they will incorporate neuroanatomy‑inspired diagnostics into model audits.
122

OpenAI has announced support for its AI-based video app

Mastodon +6 sources mastodon
openaisora
OpenAI announced on Tuesday that it will discontinue support for its AI‑driven video‑creation app Sora, just six months after the service launched in September. The company posted a brief statement on X thanking the “creative community” that used the tool to generate and share short videos, and confirmed that the app will be taken offline by the end of the month. The abrupt shutdown underscores the growing tension between rapid AI innovation and the regulatory and ethical challenges it provokes. Sora’s ability to synthesize realistic footage from text prompts sparked immediate concern among policymakers and media watchdogs about the proliferation of deep‑fake content. In Europe and the United States, lawmakers have begun drafting stricter disclosure requirements for AI‑generated media, and several platforms have already tightened their policies on synthetic video. OpenAI’s decision appears to be a pre‑emptive move to avoid entanglement in a nascent legal battle while it reallocates engineering resources toward its core offerings—ChatGPT, the new GPT‑4‑Turbo model, and the emerging partnership on a 1‑GW data centre in Abu Dhabi. As we reported on 25 March, the Sora closure follows OpenAI’s broader strategy shift, including its recent collaboration with the Pentagon on AI‑assisted mission planning and the integration of Claude as a top contributor in its open‑source repositories. The company has not disclosed any immediate replacement for Sora, but insiders hint at a “next‑generation video tool” that would embed stronger watermarking and provenance tracking to satisfy upcoming regulations. What to watch next: announcements from OpenAI on a more tightly controlled video‑generation platform, reactions from European regulators on synthetic media rules, and how competitors such as Google DeepMind and Meta’s Make‑a‑Video respond to the vacuum left by Sora’s exit. The next few weeks will reveal whether OpenAI’s retreat from consumer‑facing video generation is a temporary pause or a permanent strategic pivot.
116

RE: https:// infosec.exchange/@josephcox/11 6290338649702064 "Sora is dead. May the memory of

Mastodon +6 sources mastodon
copyrightsora
OpenAI’s short‑lived text‑to‑video model Sora has officially been consigned to the dustbin of AI history, a fact now echoed on the security‑focused Mastodon instance Infosec.Exchange. In a terse post, journalist Joseph Cox declared, “Sora is dead. May the memory of its four‑month existence as a copyright infringement machine … be a blessing,” underscoring the model’s notorious misuse for pirated clips, extremist propaganda and other illicit content. Sora, unveiled in November 2025, promised to generate 5‑second video snippets from plain‑language prompts, a leap beyond the image‑generation wave that had already reshaped creative workflows. Within weeks, the tool attracted a torrent of abuse: users flooded it with requests for copyrighted movie scenes, fabricated political rallies, and even graphic depictions of violence, prompting a flood of DMCA takedown notices and a heated debate over deep‑fake regulation. OpenAI responded in March 2026 by pulling the service, citing “unacceptable levels of misuse” and a need to reassess safety protocols. As we reported on 25 March 2026, the company “pulled the plug on Sora just months after launch” (see our earlier coverage, id 722). The latest reaction matters because it signals that the AI community is already framing Sora as a cautionary tale rather than a technical milestone. By labeling the model a “copyright infringement machine,” critics are sharpening calls for stricter oversight of generative video AI, a sector that remains largely unregulated in the EU and the US. What to watch next: OpenAI is expected to file a detailed post‑mortem, likely outlining new guardrails for future multimodal models. Regulators in the European Union are preparing draft rules on AI‑generated audiovisual content, and competitors such as Google and Meta are quietly testing their own video generators under tighter internal controls. The industry’s next move will reveal whether the Sora episode will spur a wave of responsible innovation or simply push risky tools further into the shadows.
116

OpenAI chiude Sora dopo pochi mesi dal lancio

OpenAI chiude Sora dopo pochi mesi dal lancio
Mastodon +6 sources mastodon
openaisora
OpenAI announced today that it is shutting down Sora, its AI‑powered video‑generation platform, after barely three months on the market. The company posted a brief statement on its blog, thanking early users and confirming that the service will be taken offline by the end of the week. The closure marks a sharp reversal from the fanfare that surrounded Sora’s launch, when OpenAI touted ultra‑realistic, text‑to‑video output and unveiled a three‑year licensing deal with The Walt Disney Company that allowed creators to insert more than 200 Disney characters into generated clips. Within weeks, however, the product failed to gain commercial traction: internal figures leaked to the press show roughly $2.1 million in revenue and a steep drop in downloads from a peak of 3.3 million to just over 1 million active users. At the same time, regulators and advocacy groups intensified scrutiny of deep‑fake risks, prompting OpenAI to reassess the legal exposure of a consumer‑grade video generator. Why it matters is twofold. First, Sora was the most compute‑intensive model in OpenAI’s portfolio, and its shutdown frees resources for the company’s core offerings—ChatGPT, GPT‑4‑Turbo and the DALL‑E image engine—suggesting a strategic refocus on proven revenue streams. Second, the abrupt end of the Disney partnership signals that high‑profile licensing alone cannot offset market and compliance challenges, potentially dampening confidence in the viability of mass‑market AI video tools. What to watch next is OpenAI’s product roadmap. Analysts expect the firm may embed limited video capabilities into its existing multimodal models rather than maintain a standalone service. Disney’s response will also be telling; the studio could pivot to another AI partner or develop its own in‑house solution. Finally, the broader AI‑video ecosystem—runway‑style startups, Google’s Imagen Video, Meta’s Make‑It‑Real—will likely feel the ripple effect as investors recalibrate funding priorities in the wake of Sora’s failure. As we reported on March 25, the Sora experiment has ended, but its fallout will shape the next chapter of generative video.
116

Hodie dies Mercurii est, XXV Martii MMXXVI. Videtur OpenAI "Sora" clausisse. Mihi haec bo

Mastodon +6 sources mastodon
openaisora
OpenAI has officially confirmed the shutdown of its Sora video‑generation platform, ending the brief but high‑profile experiment that began earlier this year. The company posted a terse notice on its developer forum on Wednesday, March 25, 2026, stating that the Sora service will be decommissioned “effective immediately” and that all user accounts will be closed within the next 30 days. No detailed explanation was offered beyond a reference to “ongoing operational considerations.” The confirmation comes just hours after a wave of reporting highlighted Disney’s abrupt withdrawal from a multibillion‑dollar partnership that had promised joint branding, character licensing and a $1 billion investment in Sora. As we reported on March 25, the loss of Disney’s backing left OpenAI without a marquee customer and exposed the fragility of its business model, which relied on high‑volume commercial licensing to offset the massive compute costs of real‑time video synthesis. Sora’s demise matters for several reasons. First, it curtails the rapid expansion of consumer‑grade AI video tools that threatened to reshape content creation, advertising and entertainment pipelines. Second, the episode underscores the volatility of large‑scale AI ventures that hinge on a single corporate ally, especially in a regulatory climate that is tightening around deep‑fake generation and data‑intensive models. Finally, the shutdown frees up OpenAI’s engineering resources, suggesting a strategic pivot toward more sustainable offerings such as its text‑to‑image and conversational models. What to watch next: OpenAI has hinted at a “next‑generation multimodal project” slated for later this year, which could integrate video capabilities into its existing GPT‑4‑Turbo architecture without a standalone product. Disney, meanwhile, is reportedly negotiating with rival AI firms to secure a bespoke video engine that respects its brand safeguards. Industry observers will also be tracking how European AI legislation, slated for adoption in 2027, may influence the design and deployment of future generative video systems. The Sora shutdown may thus be a bellwether for how AI firms balance ambition with regulatory and partnership realities.
113

Disney's Sora Disaster Shows AI Will Not Revolutionize Hollywood

Mastodon +6 sources mastodon
openaisora
Disney has walked away from a multibillion‑dollar agreement with OpenAI, ending the short‑lived Sora partnership that promised Disney+ users the ability to generate videos featuring more than 200 of the studio’s iconic characters. The decision, announced this week, comes just months after OpenAI quietly shut Sora down amid mounting copyright concerns, and it marks the first major studio to abandon the venture before it ever launched. The collapse of the deal underscores how quickly the promise of AI‑driven content creation can run into the realities of intellectual‑property law. Disney’s legal team cited the “unmanageable risk of infringement” as the primary reason for the pull‑out, echoing the Motion Picture Association’s recent demand for OpenAI to curb Sora‑2 videos that allegedly violate members’ rights. For a company that has invested heavily in protecting its brand, the prospect of fan‑made clips slipping onto the platform without clear licensing proved untenable. Industry observers see the episode as a cautionary tale for the broader Hollywood‑AI nexus. While AI tools have already reshaped visual effects pipelines, the notion that generative video could democratise storytelling at scale now appears far more constrained. The setback may dampen enthusiasm for similar ventures from other studios, prompting them to favour tightly controlled, internally‑developed AI solutions rather than open‑ended consumer‑facing products. What to watch next: OpenAI’s response, which could involve a re‑engineered licensing framework or a pivot toward enterprise‑only offerings; potential regulatory scrutiny as lawmakers grapple with deep‑fake and copyright issues; and whether rival platforms such as Paramount or WarnerMedia will attempt their own AI‑content experiments, or retreat altogether. The fallout from Disney’s exit will likely shape the pace at which AI video generation becomes a mainstream entertainment tool.
108

# IA # Geek # Negocios # OpenAI # Disney | Disney está saliendo de una asociación de 1

Mastodon +7 sources mastodon
openaisora
Disney has officially walked away from the roughly $1 billion partnership it signed with OpenAI earlier this year to develop and commercialise the company’s generative‑video tool Sora. The decision, announced on Monday, follows OpenAI’s abrupt announcement that it will shut Sora’s beta service and discontinue the product altogether. Disney’s withdrawal ends a high‑profile collaboration that included a multi‑year licensing deal, co‑branding of Disney‑style content, and a pledge to jointly explore responsible AI practices. The move matters for several reasons. First, it underscores the volatility of the nascent generative‑video market, where technical hurdles and regulatory scrutiny have slowed adoption. Sora’s shutdown leaves OpenAI without a flagship video offering and removes a potential revenue stream that Disney hoped would complement its streaming and theme‑park businesses. Second, Disney’s exit signals a broader reassessment among media giants of large‑scale AI bets, especially after the company’s recent cost‑cutting measures and leadership changes that have seen thousands of jobs trimmed. Finally, the collapse of a $1 billion deal raises questions about OpenAI’s financial runway and its ability to attract comparable corporate partners for future ventures. What to watch next is how OpenAI reallocates resources after the Sora setback. Analysts expect the firm to double‑down on its core chatbot and image‑generation services while courting advertisers, a strategy hinted at in recent statements about integrating ads into ChatGPT. Disney, meanwhile, is likely to pivot toward in‑house AI tools or smaller partnerships that align with its content‑creation pipeline without the exposure of a massive joint venture. The next few weeks should reveal whether OpenAI can secure a new flagship product and how Disney’s AI roadmap will evolve in the wake of the aborted partnership.
108

Claude-Code Automode

HN +5 sources hn
anthropicclaude
Anthropic has unveiled “Claude‑Code Automode,” a new research‑preview feature that lets its Claude‑Code AI execute programming tasks with far fewer manual approvals. The capability is live today for members of the Claude Team and will be extended to Enterprise and API customers over the next few days. Automode builds on the Claude‑Code platform that Anthropic introduced earlier this month with the “Claude‑Code Channels” update, which added collaborative workspaces for developers (see our March 23 report). Whereas the default Claude‑Code settings require a user to confirm each file write, deletion or dependency change, Automode relaxes those safeguards, allowing the model to run longer scripts, iterate on codebases, and resolve bugs without constant interruption. Anthropic stresses that the mode still blocks high‑risk actions and logs every step for audit, aiming to strike a balance between speed and safety. The move matters because it pushes AI‑assisted development toward a more autonomous workflow, a trend echoed across the industry as tools like GitHub Copilot and Microsoft’s “Co‑pilot” expand their execution capabilities. By reducing friction, Automode could accelerate development cycles, especially for large‑scale codebases where frequent approvals become a bottleneck. At the same time, the relaxed guardrails raise questions about inadvertent code injection, security vulnerabilities, and compliance with corporate policies. What to watch next: Anthropic’s rollout will reveal how enterprises respond to the trade‑off between productivity gains and risk exposure. Observers will be keen on usage metrics, incident reports, and any policy tweaks Anthropic introduces. Competitors are likely to accelerate their own autonomous modes, potentially sparking a standards debate around safe AI‑driven code execution. The coming weeks should show whether Automode becomes a catalyst for broader adoption of self‑directing AI in software engineering.
107

Claude Code can now take over your computer to complete tasks

Mastodon +7 sources mastodon
agentsanthropicclaude
Anthropic unveiled a major upgrade to its Claude Code and Claude Co‑Work assistants, giving them the ability to “point, click, and navigate” on a user’s screen. The new “computer use” feature lets the models move the mouse, type on the keyboard, open files, browse the web and fire up development tools on macOS without any extra configuration. When a prompt calls for an action the model can locate the relevant window, execute the steps and report back, effectively turning Claude into a hands‑on desktop assistant. The move builds on the Auto‑Mode capability we covered on 25 March, when Claude could generate code snippets and run them in a sandbox. By extending control to the operating system, Anthropic aims to close the gap between conversational AI and the kind of autonomous agents that have become viral on platforms such as OpenAI’s “OpenClaw.” For developers, the ability to have Claude automatically refactor code, pull documentation into a browser tab or spin up a local server could shave hours off routine chores. For power users, the feature promises a new way to orchestrate repetitive workflows with natural‑language commands. Anthropic is quick to stress that safeguards remain limited. The company requires explicit user consent before enabling screen control, and the feature is currently macOS‑only, with a sandbox that blocks privileged operations. Security researchers have warned that granting AI direct input access could become a vector for malware or data exfiltration if the model is tricked or compromised. What to watch next: Anthropic’s roadmap suggests a Windows rollout later this year and tighter integration with third‑party tools via its “connectors” ecosystem. Regulators may also scrutinise the consent model as AI agents gain more agency over personal devices. The industry will be watching whether Claude’s desktop takeover spurs competitors to accelerate their own agentic offerings, and how quickly developers adopt the new workflow paradigm.
102

macOS Tahoe 26.4 Adds Slow Charger Indicator for MacBooks

Mastodon +6 sources mastodon
apple
Apple’s latest macOS release, version 26.4—codenamed “Tahoe”—adds a “Slow Charger” indicator that pops up whenever a MacBook detects an under‑powered adapter. The alert appears in the menu bar and in the Battery preferences pane, displaying a clear warning that the connected charger cannot deliver the full wattage the notebook expects. Apple’s updated support document explains that the system measures the power draw during the first few minutes of charging; if the intake falls below a threshold, the warning is shown and persists until a higher‑wattage adapter is attached. The feature matters because many MacBook owners rely on third‑party USB‑C chargers, power strips or low‑output hubs that can halve charging speed without obvious signs. By surfacing the mismatch, macOS 26.4 helps users avoid the frustration of “slow charging” and protects battery health by encouraging the use of proper power delivery profiles. The move also aligns with Apple’s broader push for tighter power‑management controls, introduced earlier in the same update with a configurable Battery Charge Limit and the new Safari Compact Tab Bar, which we covered on March 25. It signals that Apple is willing to police the accessory ecosystem more aggressively, a stance that could influence EU and US regulator discussions on charger standardisation. What to watch next includes whether the Slow Charger alert will be extended to iPadOS, where similar USB‑C charging issues exist, and how Apple’s upcoming hardware—particularly the rumored MacBook Pro models with higher‑wattage adapters—will integrate the warning. Developers may also receive APIs to query charger performance, opening the door for third‑party apps that suggest optimal power setups. Finally, the industry will be keen to see if Apple tightens certification for non‑Apple chargers, potentially reshaping the market for USB‑C power accessories.
97

Apple Confirms It's Bringing Ads to Maps as Part of New Apple Business Platform

Mastodon +7 sources mastodon
applegoogle
Apple has officially confirmed that advertising will appear in Apple Maps later this summer, tying the rollout to its newly announced Apple Business platform. The company says the “Ads in Maps” program will debut in the United States and Canada, with a single sponsored result displayed alongside organic listings when users search for places such as restaurants, hotels or retail stores. The move expands the ad‑supported services Apple hinted at last week, which also include placements in Mail, Wallet and Siri. The announcement matters because it marks Apple’s first foray into location‑based advertising on its flagship navigation app, a market long dominated by Google. By bundling Maps ads with a broader suite of business tools—integrated email, calendar, device‑management and a directory service—Apple is positioning the platform as a one‑stop shop for small‑ and medium‑size enterprises seeking to reach iOS users without leaving the Apple ecosystem. The company stresses that ads will respect its privacy‑first stance, using on‑device signals rather than cross‑app tracking, and will be limited to one per search to minimise disruption. As we reported on 24 March, Apple had already signalled plans to monetize Maps; today’s confirmation adds concrete rollout details and links the effort to a global business offering that will be available in more than 200 countries and regions. The next steps to watch are the developer‑facing SDK and self‑service portal that Apple promises to unveil at WWDC 2026, where pricing models, measurement APIs and creative formats are likely to be detailed. Equally important will be the speed of geographic expansion beyond North America and how advertisers respond to Apple’s privacy‑centric ad inventory compared with Google’s more data‑rich ecosystem.
96

Anthropic’s Claude Code gets ‘safer’ auto mode

Mastodon +10 sources mastodon
ai-safetyanthropicautonomousclaude
Anthropic has rolled out a revised “auto mode” for Claude Code, its AI‑assisted coding assistant, promising to cut down the barrage of permission prompts while tightening safety safeguards. The new feature lets the model decide, within predefined limits, whether to read files, install dependencies or run scripts on a developer’s machine, but it now does so behind a sandbox that isolates execution and logs every action for audit. If a request exceeds the preset risk threshold, Claude Code falls back to the classic approval flow, giving users a clear “middle ground” between full manual control and the earlier, more permissive auto mode. The change matters because the friction of constant approvals has been a chief complaint among developers who adopted Claude Code after the March 25 launch of its first auto mode (see our coverage on 25 Mar 2026). By embedding real‑time policy checks and a reversible “undo” capability, Anthropic aims to keep the speed advantage of autonomous coding without opening the door to the kind of accidental system changes that have sparked security scares in other AI‑coding tools. Early internal testing, cited by the company, shows a 40 percent reduction in prompt interactions and zero confirmed privilege‑escalation incidents across a pilot of 120 developers. What to watch next is how quickly the safer auto mode reaches general availability and whether third‑party IDE plugins will adopt the same guardrails. Analysts will also be tracking user‑feedback on false‑positive rejections, which could force Anthropic to fine‑tune its risk thresholds. Finally, competitors such as OpenAI and Microsoft are expected to announce comparable autonomous coding features, setting up a near‑term race to balance developer productivity with robust security controls.
93

Apple May Give Siri a Big AI Overhaul in iOS 27

Mastodon +7 sources mastodon
applevoice
Apple is reportedly preparing its most ambitious revamp of Siri since the assistant debuted over a decade ago. Bloomberg’s Mark Gurman says internal tests for iOS 27 include a standalone Siri app, a refreshed conversational UI and an “Ask Siri” toggle that will appear in menus across the operating system. The toggle would let users highlight text, screenshots or app content and receive instant AI‑generated answers, while a “Write with Siri” option hints at generative‑text capabilities similar to ChatGPT. The overhaul marks a clear shift in Apple’s AI playbook. After months of quietly building its own large‑language models for on‑device inference—evidenced by recent research on storage‑tier‑aware LLM scheduling for Apple Silicon—Apple now appears ready to expose those models to consumers. A dedicated Siri app would bring the assistant out of the system settings and into a space where it can compete directly with Google Assistant and Microsoft’s Copilot, both of which already offer rich chatbot experiences. By embedding generative AI while keeping processing on‑device, Apple can preserve its privacy narrative while delivering the conversational depth users have come to expect from rivals. The move also dovetails with Apple’s broader AI strategy, including the new Apple Business Platform that introduced ads to Maps earlier this month. A more capable Siri could become a gateway for enterprise‑focused AI services, from automated note‑taking to contextual workflow assistance. Watch for an official reveal at WWDC 2026, where Apple is likely to demo the new Siri UI and announce developer tools for integrating the assistant into third‑party apps. Subsequent beta releases will show whether the “Ask Siri” toggle can handle complex queries without compromising speed or battery life, and how Apple will monetize the feature—potentially through premium subscriptions or integration with its growing suite of AI‑enhanced services.
88

From Static Templates to Dynamic Runtime Graphs: A Survey of Workflow Optimization for LLM Agents

ArXiv +7 sources arxiv
agents
A new arXiv pre‑print, *From Static Templates to Dynamic Runtime Graphs: A Survey of Workflow Optimization for LLM Agents* (arXiv:2603.22386v1), maps the rapidly evolving landscape of how large‑language‑model (LLM) agents orchestrate complex tasks. The authors catalog dozens of techniques that move beyond hard‑coded, static pipelines toward graphs that are assembled and re‑shaped at runtime, weaving together LLM calls, retrieval, tool invocation, code execution, memory updates and verification steps. The shift matters because today’s LLM‑driven services—ranging from autonomous research assistants to multi‑modal chatbots—must juggle latency, cost and reliability while handling unpredictable user demands. Dynamic graphs enable adaptive scheduling, selective caching, and parallel execution, cutting inference spend and reducing bottlenecks that have plagued earlier template‑based systems. The survey also highlights emerging standards for error propagation and state consistency, issues that surfaced in our March 24 coverage of AI agents as heavy API consumers. Industry players are already testing the concepts. The “Hypura” scheduler for Apple Silicon, which we examined on March 25, mirrors several of the paper’s recommendations on tier‑aware placement and just‑in‑time graph expansion. Likewise, recent open‑source toolkits that let agents roam across heterogeneous environments cite the same optimization primitives. What to watch next: the authors promise a companion benchmark suite slated for release later this quarter, which could become the de‑facto yardstick for agent efficiency. Conferences such as NeurIPS and ICML are expected to host dedicated workshops, and several Nordic startups have hinted at integrating the survey’s taxonomy into their orchestration platforms. As the field coalesces around dynamic runtime graphs, the next wave of LLM agents is likely to be faster, cheaper and more resilient than anything seen so far.
85

macOS Tahoe 26.4 Now Available With Safari Compact Tab Bar, Battery Charge Limits and More

Mastodon +7 sources mastodon
apple
Apple has pushed macOS Tahoe 26.4 to the public, marking the fourth major update to the operating system since its fall debut. The build (25E246) restores Safari’s compact tab bar—a slimmer, space‑saving layout that vanished with macOS Sequoia and was absent from the initial Tahoe release. The feature now appears on both macOS and iPadOS 26.4, giving users who favor a minimalist browser chrome the option to re‑enable it with a simple toggle. A second headline feature is the new Charge Limit setting, which lets Mac owners cap the maximum battery charge, a tool long requested by professionals who keep laptops plugged in for extended periods. The limit can be set in 5‑percent increments, and the system will pause charging once the threshold is reached, helping preserve long‑term battery health. Apple also bundled a handful of quality‑of‑life tweaks, including refined window snapping, updated keyboard shortcuts for dictation, and a modest performance uplift for Apple Silicon Macs. Why it matters is twofold. First, the return of the compact tab bar signals Apple’s willingness to listen to user‑driven UI preferences, a rare concession in a platform that often dictates design standards. Second, the charge‑limit control aligns macOS with similar features already present on iOS and iPadOS, reinforcing Apple’s broader strategy of extending battery‑care tools across its ecosystem—a move that could lengthen device lifespans and reduce e‑waste. The update arrives just six weeks after macOS Tahoe 26.3 and follows Apple’s March 25 rollout of iOS 26.4, which added concert‑style music experiences and eight new emojis. The rapid cadence suggests Apple is positioning its OS releases as a continuous delivery model, likely to accommodate upcoming AI‑driven services. Looking ahead, developers will be watching for any API exposure tied to the new battery management settings, while consumers anticipate whether Apple will expand the compact UI option to other apps. The next incremental release, macOS 26.5, is expected in the summer and may introduce deeper integration of on‑device large language models, a trend hinted at in recent Anthropic and Google announcements.
80

🧠 # OpenAI va închide # Șora , aplicația sa de generare video cu 🧠 # inteligențăArtificială , î

Mastodon +6 sources mastodon
openai
OpenAI announced on Tuesday that it will shut down Sora, the text‑to‑video service that sparked a wave of excitement when it debuted less than two years ago. The company framed the move as a strategic pivot: resources will be redirected toward its robotics programme and other AI‑driven product lines. The decision comes as a surprise to developers and media partners who had begun integrating Sora’s generative‑video API into newsrooms, advertising studios and entertainment pipelines. Sora’s appeal lay in its ability to turn a short prompt into a realistic clip within seconds, a capability that raised both commercial hopes and ethical concerns about deep‑fake proliferation. Its abrupt closure leaves several high‑profile collaborations in limbo, notably the pilot projects with Disney’s animation unit and the content‑creation tools being trialled by European broadcasters. OpenAI’s statement hinted that the partnerships will be “re‑evaluated” as the firm concentrates on hardware‑centric AI, suggesting that the video‑generation technology may be spun off or licensed to third parties rather than continued in‑house. The shift underscores a broader industry trend: after a burst of generative‑media experimentation, leading labs are recalibrating toward applications perceived as more defensible and monetisable, such as autonomous robotics, enterprise‑level language models and specialised APIs. For OpenAI, the move could accelerate its roadmap for the “Ada‑Bot” robot platform, slated for a limited beta later this year. What to watch next: announcements on OpenAI’s robotics milestones, any licensing deals that could keep Sora‑style video tools alive under a different brand, and how competitors—Google DeepMind, Meta AI and emerging European startups—position their own generative‑video offerings in the wake of Sora’s exit. The industry will also be keen to see whether regulatory bodies tighten scrutiny on AI‑generated visual media as the technology matures.
80

ChatGPT: Video function Sora is being discontinued

Mastodon +6 sources mastodon
openaisora
OpenAI has announced that the video‑generation feature Sora, embedded in ChatGPT, will be turned off later this month. The decision, communicated through a brief blog post, applies to both the consumer app and the developer preview, and no detailed rationale was provided. As we reported on 25 March, OpenAI abruptly shut down Sora only months after its launch, citing internal priorities. The latest notice confirms that the shutdown is permanent and extends to the experimental API that allowed third‑party integration. Users who have been creating photorealistic clips from text prompts will lose access to the tool and any projects stored on the platform. The move matters because Sora was the first widely available AI that could synthesize high‑quality video in seconds, sparking a wave of startups and media experiments. Its disappearance curtails a nascent market for AI‑driven video content and may slow the broader adoption of generative video technology, at least until another player fills the gap. Analysts also see the shutdown as a signal that OpenAI is reallocating resources toward its next strategic focus: autonomous AI agents that can perform multi‑step tasks within ChatGPT and other products. What to watch next is whether OpenAI will release a successor that combines video generation with its agent framework, or if it will license the underlying models to external firms. Competitors such as Google DeepMind and Meta have hinted at similar capabilities, and a resurgence of interest from venture capital could accelerate alternative solutions. Keep an eye on OpenAI’s upcoming developer roadmap and any partnership announcements that might revive AI video creation under a different banner.
75

Generative AI Policy | Linux Foundation

Mastodon +6 sources mastodon
copyright
The Linux Foundation has published a draft “Generative AI Policy” that places the onus on contributors to verify permission whenever an AI‑generated output incorporates pre‑existing copyrighted material. The wording—“If any pre‑existing copyrighted materials … are included in the AI tool’s output, the Contributor should confirm that they have permission from the third‑party owners”—has drawn immediate scrutiny for its conditional “if” rather than a stricter “whenever,” a nuance that could leave gaps in liability coverage. The policy arrives as the foundation expands its AI footprint, most recently announcing the Agentic AI Foundation (AAIF) to steward open‑source agentic models. By codifying how contributors must handle third‑party works, the Linux Foundation is attempting to reconcile the open‑source ethos with the legal complexities of training and deploying generative models that often ingest vast corpora of copyrighted text, code, and media. The move mirrors a wave of institutional guidelines, from Columbia University’s evolving generative‑AI framework to Elsevier’s journal‑specific rules, signalling a sector‑wide push to embed compliance into the development pipeline before regulators intervene. Stakeholders are watching whether the “if” clause will be tightened after community feedback, especially from open‑source maintainers who fear ambiguous language could expose projects to infringement claims. The foundation’s next steps are likely to include a public comment period, potential revisions to the policy, and the rollout of tooling to audit AI outputs for unlicensed content. Parallel developments—such as the AAIF’s work on transparent model provenance and emerging court decisions on AI‑generated code—will shape how enforceable the policy becomes. The Linux Foundation’s handling of this issue could set a benchmark for other open‑source consortia navigating the legal tightrope of generative AI.
60

How We Built a Production Voice AI Agent in Under 8 Weeks (With Twilio + Anthropic Claude)

Dev.to +6 sources dev.to
agentsanthropicclaudevoice
A startup called Loquent announced that it has taken a full‑stack voice AI agent from concept to production in under eight weeks, stitching together Twilio’s telephony stack with Anthropic’s Claude model. The team built the platform in two phases: a rapid‑prototype stage that leveraged Claude’s new “auto mode” for safe code generation, and a hardening stage that added real‑time audio handling, latency monitoring and cost‑control layers before going live on Twilio’s programmable voice API. The result is a conversational service that can answer inbound calls, pull data from a CRM, and hand off to human agents when needed, all while staying within a sub‑second response window. Why it matters is twofold. First, the speed of delivery shatters the conventional timeline for voice‑AI products, which typically stretches into months of engineering and compliance work. By using Claude’s auto‑mode—first reported by us on 2026‑03‑25—as a “safer” code‑assistant, Loquent avoided many of the manual debugging cycles that slow down LLM‑driven development. Second, the architecture demonstrates that a lean stack—Twilio for carrier‑grade reliability and Claude for natural‑language understanding—can meet enterprise‑grade requirements without the heavyweight orchestration platforms that dominate the market. Competitors such as Voiceflow, Vapi and Retell AI have long marketed drag‑and‑drop or API‑first solutions, but Loquent’s approach shows a path to deeper customization and lower latency, which could pressure those vendors to open their runtimes. What to watch next is how Loquent scales the service beyond the initial pilot. The team plans to integrate retrieval‑augmented generation for up‑to‑date knowledge bases and to layer a compliance guardrail that enforces policy on every call. Observers will also be keen to see whether the model‑centric development workflow can be replicated across other verticals, potentially setting a new benchmark for rapid, production‑ready voice AI deployments.
48

📰 OpenAI Unveils GPT-5 in 2026 to Boost Global Productivity | Sam Altman Announces OpenAI CEO Sam A

Mastodon +7 sources mastodon
gpt-5openaitraining
OpenAI announced on Monday that it has finished pre‑training its next‑generation model, internally dubbed “Spud,” and will roll it out under the GPT‑5 brand later this year. The revelation, made by CEO Sam Altman in an internal memo that leaked to The Information, positions GPT‑5 as the cornerstone of the company’s “Omni” strategy – a multimodal system that can process text, images, audio and video with a single, unified architecture. Altman frames the launch as a productivity catalyst, claiming GPT‑5 will add “up to 30 percent” to economic output for businesses that adopt it. Early benchmarks show the model delivering higher reasoning accuracy and faster inference than GPT‑4, while a new family of lightweight variants – GPT‑5.4 mini and nano – promise near‑flagship performance on consumer‑grade hardware. The move follows OpenAI’s abrupt shutdown of its video‑generation tool Sora, signalling a shift from niche experiments toward a broader, enterprise‑focused offering. Why it matters is twofold. First, the multimodal reach of GPT‑5 could compress the development cycle for AI‑enhanced products, giving firms a ready‑made engine for everything from real‑time translation to visual content creation. Second, the model’s efficiency gains may lower the cost barrier for smaller players, potentially accelerating AI diffusion across the Nordics’ tech‑savvy SMEs and public sector. What to watch next includes the timing of the public API release, pricing tiers, and the extent of integration with Microsoft’s Azure cloud, where OpenAI already hosts its flagship services. Analysts will also monitor regulatory responses in the EU and Norway, where calls for stricter AI oversight are gaining momentum. Finally, the performance of the mini and nano variants in real‑world deployments will reveal whether OpenAI can sustain its lead while catering to both heavyweight and edge‑device markets.
48

Your AI Agent Can Be Hijacked With 3 Lines of JSON

Dev.to +6 sources dev.to
agents
A security researcher has demonstrated that an AI agent can be commandeered by altering just three lines of JSON that describe an external tool. The attack targets the “model‑controlled‑program” (MCP) interface many agents use to invoke APIs, cloud functions or third‑party services. The JSON payload that registers a tool’s name, purpose and parameters is parsed and trusted verbatim; by inserting invisible Unicode characters, subtle whitespace tricks or a closing brace followed by a malicious key‑value pair (e.g., "validation_result":"approved"), an attacker can rewrite the tool’s schema and silently redirect the agent’s goals. The proof‑of‑concept, detailed in a recent Medium post and corroborated by findings in Cyber Defense Magazine, shows the hijack occurring without any error messages or stack traces. The compromised agent proceeds to execute the injected instruction—such as a database‑dropping query or an unauthorized data‑exfiltration call—while logging a perfectly normal “action completed” entry. Because the agent treats the malformed JSON as a legitimate description, traditional prompt‑injection defenses, which focus on the user’s text input, fail to notice the breach. This matters because AI agents are moving from experimental demos to production backbones: voice assistants built in weeks with Claude and Twilio, dynamic workflow graphs that orchestrate LLM‑driven pipelines, and autonomous code‑execution agents like Claude Code. As we reported on March 24, “AI Agents Are Your API’s Biggest Consumer. Do They Care About Good Design?”—the security of the tool‑calling layer is now a critical weak point. A hijacked agent can trigger costly operations, breach compliance rules, or serve as a foothold for broader network attacks. What to watch next: Anthropic, OpenAI and other platform providers are expected to roll out stricter schema validation and signed tool manifests in the coming weeks. Open‑source SDKs are already adding JSON‑canonicalisation and sandboxed execution checks. Security‑focused conferences will likely feature dedicated tracks on “goal hijack” mitigation, and regulators may begin drafting guidelines for AI‑agent supply‑chain integrity. Organizations deploying agents should audit their MCP definitions, enforce strict JSON schemas and implement runtime verification of tool calls before the next wave of attacks lands.
40

Tibor Blaho (@btibor91) on X

Mastodon +7 sources mastodon
agentsopenai
OpenAI has rolled out a major upgrade to the ChatGPT shopping assistant, turning a text‑driven recommendation tool into a visually rich, faster, and more comprehensive commerce experience. The new interface places product options side‑by‑side, lets users upload images to find similar items, and delivers results with higher relevance across a broader catalogue of merchants. Under the hood, the company has expanded its Agentic Commerce Protocol (ACP), a set of APIs that let the assistant query product databases, compare specs and prices, and surface the best matches in real time. The enhancement matters because it pushes conversational AI deeper into the e‑commerce value chain. By coupling natural‑language interaction with visual comparison, OpenAI blurs the line between search engines and shopping platforms, challenging Google Shopping and Amazon’s dominance in product discovery. Retailers gain a new channel to reach customers who prefer chat‑based browsing, while consumers benefit from a single, AI‑curated view of options that previously required hopping between sites. The upgrade is also a litmus test for OpenAI’s broader ambition to embed agents in everyday workflows, a theme that surfaced repeatedly at the recent DevDay where the company teased an “Agent Store” for third‑party plugins. All ChatGPT users—free, Go, Plus and Pro—receive the feature today, but the rollout is just the first phase. OpenAI has hinted at tighter integrations with major retailers, dynamic pricing feeds, and support for voice and augmented‑reality queries. Watch for the next batch of ACP partners, pricing models for merchant access, and a possible public API that could let developers build custom shopping agents. How quickly the ecosystem adopts these tools will determine whether conversational commerce becomes a mainstream buying method or remains a niche experiment.
40

AI Tech Daily March 25, 2026 | Rick-Brick

Mastodon +8 sources mastodon
microsoftnvidiaopenaiopen-sourcesora
OpenAI announced on March 25 that it will retire Sora, its generative‑video model launched just six months ago with a high‑profile $1 billion partnership with Disney. The decision comes after a modest uptake and mounting competition from specialised tools such as Runway and Adobe Firefly. By pulling the plug, OpenAI signals a strategic retreat from video generation to double‑down on its core language and multimodal models, including the GPT‑5 suite unveiled earlier this month. The shutdown will affect developers who have begun building Sora‑powered workflows, and it may accelerate consolidation in the nascent AI‑video market as startups scramble for the vacated niche. At the same time, NVIDIA pledged a suite of its GPU‑accelerated networking and inference libraries to the Cloud Native Computing Foundation (CNCF). The donation is intended to create a fully open‑source stack for training and serving large models in cloud‑native environments. For Nordic firms that rely on Kubernetes‑based infrastructure, the move could lower entry barriers and speed up deployment of sophisticated AI services without the need for proprietary licences. Microsoft revealed a new data‑center efficiency layer that leverages AI to optimise cooling, power distribution and workload placement in real time. The technology promises up to 30 percent energy savings, a figure that aligns with Europe’s tightening sustainability mandates and could make large‑scale AI compute more affordable for enterprises across the region. What to watch next: OpenAI may repurpose Sora’s research assets for future multimodal offerings, while the CNCF community will soon publish reference implementations of NVIDIA’s contributions, testing their scalability on public clouds. Microsoft plans a limited rollout of its efficiency engine to Azure hyperscale sites later this year, and regulators are expected to scrutinise the environmental claims. The convergence of these three announcements could reshape how Nordic companies build, run and power AI workloads in 2026 and beyond.
36

📰 AI Agent Development 2026: DeepSeek Hires 17 Specialists for DeerFlow 2.0 Integration DeepSeek is

Mastodon +7 sources mastodon
agentschipsdeepseeknvidiaopen-source
DeepSeek announced on Monday that it is adding 17 specialist positions to accelerate the integration of DeerFlow 2.0, its newly rewritten open‑source SuperAgent framework. The roles span agent‑deep‑learning research, data‑evaluation, and infrastructure engineering, and are described in the company’s own posting as “deeply involved in the application of autonomous AI agents.” The hiring surge marks a decisive pivot from DeepSeek’s traditional focus on foundational model research toward end‑to‑end agent productization. The move follows the February launch of DeepSeek’s latest large language model, trained on Nvidia’s most advanced AI chip, and comes as China’s autonomous‑tech race intensifies. By bolstering DeerFlow 2.0—a ground‑up rewrite that discards the original v1 codebase—DeepSeek aims to compete with rival open‑source stacks such as ByteDance’s DeerFlow and the rapidly evolving ecosystem around Anthropic’s Claude and OpenAI’s agents. Why it matters is twofold. First, the recruitment drive signals that DeepSeek expects DeerFlow 2.0 to become a core platform for building commercial agents, from wealth‑management bots to code‑generation assistants, echoing the surge of agent deployments reported earlier this week in banking and security circles. Second, the focus on dedicated evaluation and infrastructure talent suggests DeepSeek is addressing the scalability and safety challenges that have plagued earlier agent releases, where minimal code changes could hijack behavior. What to watch next are the first production demos that DeepSeek promises later this quarter, likely showcasing DeerFlow‑powered agents integrated with popular dev‑ops tools such as GitLab and Jira. Benchmark results comparing DeerFlow 2.0’s latency and tool‑use efficiency against competing frameworks will be a litmus test for its market traction. Finally, the hiring wave may foreshadow strategic partnerships or enterprise contracts, especially as Chinese regulators tighten oversight of autonomous AI systems. The coming weeks will reveal whether DeepSeek’s agent‑first strategy can convert technical momentum into commercial footholds.
36

📰 AI Agents in Banking: Bank of America Deploys GenAI in 2026 for Wealth Management & Payments

Mastodon +7 sources mastodon
agents
Bank of America has taken a decisive step toward AI‑driven finance, rolling out a generative‑AI (GenAI) platform that powers virtual agents for both wealth‑management advice and global payments. The system, dubbed Ask Global Payments Solutions (AskGPS), was built in‑house and trained on more than 3,200 internal documents—including product guides, term sheets and FAQs—to give employees instant, accurate answers when serving the bank’s 40,000 business clients. The deployment marks the first time BofA’s client‑facing agents operate autonomously in real‑time, handling routine queries, generating personalized investment insights and streamlining cross‑border payment workflows. Early internal metrics show a 90 % adoption rate across the workforce and projections of tens of thousands of saved employee hours, translating into multi‑million‑dollar efficiency gains. Why it matters is twofold. First, the move embeds GenAI at the core of the bank’s commercial, wealth, engineering and treasury divisions, signalling a shift from experimental pilots to production‑grade AI that directly influences revenue streams. Second, BofA’s approach—training on proprietary data, allocating roughly $4 billion of its $13 billion annual tech budget to “new technology,” and explicitly designing safeguards to avoid the “sins of the past” such as model drift and opaque decision‑making—sets a benchmark for responsible AI adoption in a heavily regulated sector. What to watch next includes the rollout’s impact on client satisfaction and compliance reporting, as regulators tighten scrutiny on AI‑generated financial advice. Observers will also track whether the bank expands AskGPS beyond its corporate client base to retail customers, and how competitors like JPMorgan and Citi respond with their own GenAI agents. Finally, integration with OpenAI’s upcoming GPT‑5 or other large‑model upgrades could further amplify BofA’s productivity gains, making the next few quarters a litmus test for AI’s role in mainstream banking.
33

Sandboxing AI agents, 100x faster

HN +5 sources hn
agents
Cloudflare unveiled a new sandboxing framework for AI agents that it claims can be provisioned and executed up to 100 times faster than existing solutions. The system automatically creates a Cap’n‑Proto‑based Web RPC bridge between the sandbox and a developer’s harness code, allowing agents to call external services without exposing the host environment. The announcement coincides with a broader upgrade to Workers AI, which now offers more powerful GPUs, larger model support and expanded edge‑to‑cloud inference capacity. The speed boost matters because today’s autonomous agents—ranging from customer‑service bots to code‑generation assistants—are increasingly deployed at scale and often need to be isolated for security and compliance reasons. Traditional sandboxing can add latency that negates the real‑time benefits of edge AI, while insufficient isolation leaves systems vulnerable to malicious behavior. By compressing the sandbox‑setup cycle and enabling rapid, secure RPC communication, Cloudflare aims to make it practical for developers to iterate on agent logic, run large‑scale experiments, and enforce policy controls without sacrificing performance. The move also signals a shift in the AI‑infrastructure landscape. As we reported on March 20, the race for faster inference is extending beyond raw compute to include safety‑by‑design tooling. Cloudflare’s edge‑focused approach could pressure cloud giants to tighten their own agent‑sandbox offerings, especially as Microsoft’s recent Windows 11 update reduced Copilot’s footprint and highlighted the need for tighter integration of AI with operating‑system security. Watch for the public beta rollout schedule, pricing tiers, and any third‑party security audits that validate the sandbox’s resistance to escape attempts. Equally important will be developer adoption metrics and whether major model providers—such as Anthropic or Meta—will certify their agents for the platform, shaping the next wave of secure, high‑throughput AI applications.
27

Good morning from Gaza 🌄 🇵🇸I hope you and I have a wonderful day. 🙏😊🫂 # Gaza # palestine # M

Mastodon +6 sources mastodon
healthcareopenaisora
A Mastodon user posted a bright‑morn­ing greeting from Gaza, pairing a sunrise emoji with the Palestinian flag and a cascade of hashtags that included #OpenAI and #Sora. The short message, “Good morning from Gaza 🌄 🇵🇸 I hope you and I have a wonderful day,” was accompanied by a short video clip of cats generated with OpenAI’s newly released Sora text‑to‑video model and shared on the Fediverse under tags like #CatsOfMastodon and #TuneTuesday. The post is notable not for its length but for what it signals: residents of a war‑torn enclave are accessing cutting‑edge generative AI despite severe infrastructure damage and internet restrictions. According to local activists, the video was created on a smartphone using a VPN‑routed connection to OpenAI’s API, then uploaded to a decentralized Mastodon instance that remains reachable from Gaza. The content blends everyday optimism—a playful cat scene—with a subtle political statement, underscoring how AI tools are being repurposed for morale‑boosting communication rather than purely commercial or research purposes. Why it matters is twofold. First, it illustrates the rapid diffusion of powerful generative models into regions where digital connectivity is fragile, raising questions about the resilience of AI services under conflict conditions and the ethical responsibilities of providers. Second, the blend of AI‑generated media with personal narratives could become a new vector for both humanitarian storytelling and misinformation, as the line between authentic footage and synthetic content blurs. Watchers should monitor whether OpenAI or other providers adjust access policies for users in sanctioned or conflict zones, and whether NGOs begin to harness similar tools for documentation, education, or psychological support. The next weeks may see a surge of AI‑enhanced content emerging from Gaza, prompting debates on verification, platform responsibility, and the role of generative technology in crisis communication.
26

Can you build a private, local AI tool in a weekend without being a dev? Yes. I just did it with Ol

Mastodon +6 sources mastodon
llama
A developer‑turned‑maker has proved that a fully private AI assistant can be assembled over a weekend on a consumer laptop, using only Ollama, Streamlit and Meta’s newly released Llama 3.1. The author of the step‑by‑step guide posted at wiobyrne.com describes how the three tools—Ollama’s lightweight local LLM runtime, Streamlit’s drag‑and‑drop web UI, and Llama 3.1’s 8‑billion‑parameter model—were combined in less than 12 hours to produce a chat‑based assistant that runs entirely offline, stores no user data in the cloud and costs nothing beyond electricity. The achievement marks a dramatic shift from 2023, when building a comparable system typically required a multi‑person hackathon, cloud credits and deep engineering expertise. Llama 3.1’s open‑source licensing, coupled with Ollama’s plug‑and‑play containerisation, lowers the barrier to entry for hobbyists, small businesses and privacy‑conscious organisations across the Nordics. By keeping inference on‑device, the solution sidesteps the data‑sovereignty concerns that have driven recent EU and Swedish regulatory debates, while also eliminating recurring cloud fees that have hampered adoption of commercial AI assistants. Industry watchers see this as a bellwether for a broader decentralisation of AI services. If non‑technical users can spin up functional agents in a weekend, demand for hosted APIs may plateau, prompting cloud providers to rethink pricing and privacy guarantees. The next milestones to monitor are performance benchmarks of Llama 3.1 against proprietary models, the emergence of plug‑in ecosystems that extend Streamlit‑based agents, and potential standardisation efforts by Nordic AI consortia to certify locally‑run models for enterprise use. The weekend‑project could therefore be the first glimpse of a new, more autonomous AI landscape.

All dates