Welcome back
Curated from 200+ sources across AI & machine learning

The company said the model reduces hallucination in sensitive areas such as law, medicine, and finance, while maintaining the low latency of its predecessor.



Just two months ago, researchers at the Data Intelligence Lab at the University of Hong Kong introduced CLI-Anything, a new state-of-the-art tool that analyzes any repo’s source code and generates a structured command line interface (CLI) that AI coding agents can operate with a single command. Claude Code, Codex, OpenClaw, Cursor, and GitHub Copilot CLI are all supported, and since its launch in March, CLI‑Anything has climbed to more than 30,000 GitHub stars. But the same mechanism that makes software agent-native opens the door to agent-level poisoning. The attack community is already discussing the implications on X and security forums, translating CLI-Anything's architecture into offensive playbooks. The security problem is not what CLI-Anything does. It is what CLI-Anything represents. CLI-Anything generates SKILL.md files, the same instruction-layer artifacts that Snyk’s ToxicSkills research found laced with 76 confirmed malicious payloads across ClawHub and skills.sh in Feb

OpenAI updated the default model for ChatGPT to its new GPT-5.5 Instant, along with a new memory capability that finally shows which context shaped responses — at least some of them. This limitation signals that models are starting to create a second, incomplete memory observability layer that could conflict with existing audit systems and agent logs. GPT-5.5 Instant replaces GPT-5.3 Instant as the default ChatGPT model and is a version of its new flagship GPT-5.5 LLM. It’s supposed to be more dependable, accurate and smarter than 5.3. But it’s the introduction of memory sources, which will be enabled across all models in the platform, that could help enterprises in their projects. “When a response is personalized, you can see what context was used, such as saved memories or past chats, and delete or correct it if something is outdated or no longer relevant,” OpenAI said in a blog post. When a user asks ChatGPT something, users can tap the sources button (at the bottom of the resp

A little-known Miami-based startup called Subquadratic emerged from stealth on Tuesday with a sweeping claim: that it has built the first large language model to fully escape the mathematical constraint that has defined — and limited — every major AI system since 2017. The company claims its first model, SubQ 1M-Preview, is the first LLM built on a fully subquadratic architecture — one where compute grows linearly with context length. If that claim holds, it would be a genuine inflection point in how AI systems scale. At 12 million tokens, the company says, its architecture reduces attention compute by almost 1,000 times compared to other frontier models — a figure that, if validated independently, would dwarf the efficiency gains of any existing approach. The company is also launching three products into private beta: an API exposing the full context window, a command-line coding agent called SubQ Code, and a search tool called SubQ Search. It has raised $29 million in seed funding f

OpenAI's newest default model for ChatGPT might not make stuff up as much. Hallucinations have been an ongoing problem for AI models, but OpenAI says its new GPT-5.5 Instant model has "significant improvements in factuality across the board." The company claims that, based on "internal evaluations," GPT-5.5 Instant produced "52.5% fewer hallucinated claims" than its Instant model for GPT-5.3 "on high-stakes prompts covering areas like medicine, law, and finance." GPT-5.5 Instant also "reduced inaccurate claims by 37.3% on especially challenging conversations users had flagged for factual errors." (OpenAI has some information about how it ev … Read the full story at The Verge.

The Seattle-based startup's Series A round was led by Glilot Capital, NFX, and SignalFire, TechCrunch has exclusively learned.

Etsy's new native app within ChatGPT aims to be a conversational shopping experience for users.

ServiceNow (NYSE:NOW) used its Knowledge 2026 event to introduce major AI platform advances, including autonomous agents and an upgraded AI Control Tower for enterprise governance. The company launched ServiceNow Otto, a unified conversational AI and workflow automation experience intended to sit across the enterprise. New and expanded partnerships were announced with NVIDIA, Microsoft, Lenovo, and FedEx, focused on desktop AI agents, cross cloud governance, device lifecycle automation, and...

Elsevier, Cengage, Hachette Book Group, Macmillan and McGraw Hill are suing the Meta CEO accuses the tech giant of using their books to train its AI language system Llama.

Google's smart home ecosystem is getting its biggest update since the AI-fueled 2025 revamp.
I think 80% of UGC agencies will have to re-adapt their whole workflow in 18 months to survive and most people don't see it coming. I run a small consulting business of generating AI ads and last week I lost a $500 retainer to a client who told me they're "just doing it themselves with Claude now." I was annoyed and decided to try it. Connected Higgsfield's MCP to Claude, gave it a brief I'd normally charge $150 to scope out, and watched Claude queue 10 UGC variations on its own and after getting used to it I was shocked. Different hooks, different angles, one consistent face across all of them, reduced time. The part that broke my brain wasn't the quality. Quality is fine overall, not magic. Actually in the beginning it was generating me some trash videos, which kinda disappointed me a little. The thing that broke me was watching Claude reject its own takes, search how to make an add without me. It generated a clip, decided the framing was off, queued another one. Nobody told it to.
Anthropic's alignment team published a paper this week called Model Spec Midtraining (MSM) and I think it's one of the more practically interesting alignment results I've seen in a while. The core problem they're solving: Current alignment fine-tuning can fail to generalize. You train a model to behave well on your demonstration dataset, but put it in a novel situation and it might blackmail someone, leak data, or "alignment fake" (pretend to be aligned while actually pursuing different goals). This isn't theoretical multiple papers in 2024 documented real instances of this in LLM agents. What MSM actually does: Before fine-tuning, they add a new training stage where the model reads a diverse corpus of synthetic documents discussing its own Model Spec (the document that describes intended behavior). The idea is intuitive: instead of just showing the model what to do, you teach it why those behaviors are the right ones. Then when fine-tuning comes, the model generalizes from princip
we started analyzing features and impact take a look here https://robotics.cantarollm.tech https://preview.redd.it/c1j6imxkiczg1.png?width=370&format=png&auto=webp&s=54441cda5ca36f86b1fd9dcd63c9decfd11c8e56 send us your proposal to include new robots in the list for free submitted by /u/Careful-Newt8486 [link] [comments]

A DeepMind engineer saw an opening in SAP's aging code. Investors are calling it a $40 million bet
OpenAI is swapping out ChatGPT's default model for GPT-5.5 Instant. In internal testing, the update produced 52.5 percent fewer hallucinated claims on high-risk topics like medicine and law. A new feature called "memory sources" lets users see which stored context shaped a given response. The model is rolling out to all ChatGPT users right away, though personalization based on past chats, files, and Gmail launches first for Plus and Pro users on the web. The article ChatGPT update rolls out GPT-5.5 Instant with fewer hallucinations and more personalized answers appeared first on The Decoder.

With new AI agents, a massive joint venture, and a first-ever Jamie Dimon summit, the Claude maker is executing a two-track financial services strategy.
AI news from 200+ sources
Get Started Free
OpenAI’s president wrapped his testimony on Tuesday by revealing a fiery meeting with Musk and subsequent efforts to remove several board members.

Altara’s AI aims to diagnose failures and help speed up R&D by unifying data siloed across spreadsheets and legacy systems.

SAP plans to buy German AI startup Prior Labs and invest heavily in it. It is also prohibiting customers' agents use to a select few like Nvidia's NemoClaw.

Advanced Micro Devices (AMD) reported stronger-than-expected first-quarter results as demand for art
OpenAI is reportedly planning its own AI smartphone, with chips from MediaTek and Qualcomm and manufacturing by Luxshare. According to analyst Ming-Chi Kuo, mass production could start as early as the first half of 2027, with up to 30 million devices shipped in the first two years. The form factor choice is also an admission that more experimental AI hardware isn't ready for the mainstream yet. The article OpenAI's first hardware play might be a phone that replaces your app grid with an agent task stream appeared first on The Decoder.
The US Department of Commerce is expanding its AI safety testing: Following Anthropic and OpenAI, Google Deepmind, Microsoft, and xAI have now signed agreements with the Center for AI Standards and Innovation. The companies provide models with reduced safety guardrails for testing in classified environments amid growing cybersecurity risks and an intensifying tech race with China. The article US government now has pre-release access to AI models from five major labs for national security testing appeared first on The Decoder.

Google DeepMind, Microsoft, and Elon Musk's xAI have agreed to allow the US government to review new AI models before they're released to the public. In an announcement on Tuesday, the Commerce Department's Center for AI Standards and Innovation (CAISI) says it will work with the AI companies to perform "pre-deployment evaluations and targeted research to better assess frontier AI capabilities." CAISI, which started evaluating models from OpenAI and Anthropic in 2024, says it has performed 40 reviews so far. Both companies "have renegotiated their existing partnerships with the center to better align with priorities in President Donald Trum … Read the full story at The Verge.

With $500 million already processed through its platform, Vori is scaling fast as it targets underserved grocery operators.

Palantir Q1 revenue jumped 85% YoY to $1.63B as the firm raised its FY26 outlook to $7.66B amid surging US AI demand.

Microsoft (NasdaqGS:MSFT), Nscale, and Start Campus are launching a multi year AI data center expansion in Portugal and the EU. The project includes more than 66,000 NVIDIA Rubin GPUs and builds on an existing deployment of over 12,600 NVIDIA Blackwell Ultra GPUs for Microsoft in Portugal. The new infrastructure is described as one of Europe's largest AI data center projects and is intended to support AI workloads and cloud projects across the region. For anyone tracking Microsoft...

New AI-enabled software aims to identify small drone threats earlier and support faster response Teledyne FLIR OEM, part of Teledyne Technologies Incorporated, has announced the launch of Prism C-UAS, a new software solution designed to improve the detection and tracking of small, hard-to-detect drones. The company positions Prism as a software layer that enhances existing […] The post Teledyne FLIR OEM Launches Prism Software to Detect Small Drones at Longer Range appeared first on DRONELIFE.

Loggerhead Ventures leads the round with follow-on from One Planet Capital, fueling AI-powered drones for wind farm maintenance. Perceptual Robotics has secured fresh funding to scale its autonomous wind turbine inspection platform. Loggerhead Ventures led the round, with follow-on investment from One Planet Capital. The Thessaloniki-based company announced the deal on May 5, 2026. The […] The post Perceptual Robotics Raises Funding to Automate Wind Turbine Inspection Drones appeared first on DRONELIFE.