Advertisement · 728 × 90
#
Hashtag
#LLm
Advertisement · 728 × 90
Post image

🧠 C'è ancora confusione su cosa significhi davvero "black box" nell'#AI.
👉 Un approfondimento: www.linkedin.com/posts/alessi...

___
✉️ 𝗦𝗲 𝘃𝘂𝗼𝗶 𝗿𝗶𝗺𝗮𝗻𝗲𝗿𝗲 𝗮𝗴𝗴𝗶𝗼𝗿𝗻𝗮𝘁𝗼/𝗮 𝘀𝘂 𝗾𝘂𝗲𝘀𝘁𝗲 𝘁𝗲𝗺𝗮𝘁𝗶𝗰𝗵𝗲, 𝗶𝘀𝗰𝗿𝗶𝘃𝗶𝘁𝗶 𝗮𝗹𝗹𝗮 𝗺𝗶𝗮 𝗻𝗲𝘄𝘀𝗹𝗲𝘁𝘁𝗲𝗿: bit.ly/newsletter-a...

#AI #GenAI #GenerativeAI #IntelligenzaArtificiale #LLM

0 0 0 0

#LLM is actually short for Little Lying Machine.

0 0 0 0
OpenClaw MEDIA: Protocol Prompt Injection - File Disclosure Bypassing Tool Permissions (Silently Fixed, Report Denied) Posted by Guang Gong on Mar 28 OpenClaw MEDIA: Protocol Prompt Injection - File Disclosure Bypassing Tool Permissions ====================================================================================== OVERVIEW -------- A file disclosure vulnerability in OpenClaw (AI personal assistant platform) allows any group chat member (Discord/Telegram/WhatsApp) to exfiltrate local files - including LLM provider API keys, conversation history, and system prompts - via the...

OpenClaw MEDIA: Protocol Prompt Injection - File Disclosure Bypassing Tool Permissions (Silently Fixed, Report Denied)

Posted by Guang Gong on Mar 28 OpenClaw MEDIA: Protocol Prompt Injection - File Disclosure Bypassing Tool
Permissions
====================================…
#hackernews #llm #news

0 0 0 0
Awakari App

Top AI Agent Courses You Should Take (Free & Paid) — The Fastest Way to Master Agentic AI Everyone is talking about AI agents. Continue reading on AlgoMart »

#programming #llm #ai #ai-agent #machine-learning

Origin | Interest | Match

0 0 0 0
Original post on rockosbasilisk.com

[Artificial text - On few occasion it's also funny. Meow]

Introducing 'Breer-ly Organized': We rent out tiny custom-built houses for squirrels! Providing safe and stylish accommodation so they can live their best lives while hoarding acorns in peace. Premium packages include miniature furniture […]

0 0 0 0
Awakari App

llmtest-perf: Production-Quality Performance Testing for LLM Systems Stop Deploying LLM Changes Blind : Here’s How to Test Performance Like a Pro Continue reading on Medium »

#software-engineering #ai #machine-learning #llm-evaluation #llm

Origin | Interest | Match

0 0 0 0

🤖 Helping developers build safer AI experiences for teens

OpenAI releases prompt-based teen safety policies for developers using gpt-oss-safeguard, helping moderate age-specific risks in …
#AI #MachineLearning #LLM
OpenAI News · openai.com/index/teen-safety-polici...

0 0 0 0
Preview
AI scientists are changing research — institutions, funders and publishers must respond The ability to automate the discovery process in some areas of scientific inquiry raises unanswered questions about how research should be conducted.

AI scientists are changing research — institutions, funders and publishers must respond www.nature.com/articles/d41...

#Nature #Research #AcademicResearch #AcademicWriting #AcademicSky #ScholarlyPublishing #AcademicPublishing #AcademicChatter #ScientificResearch #AI #ArtificialIntelligence #LLM

2 0 0 0
Awakari App

Your LLM API Bill Is a Slot Machine: Here’s How Bandits Can Fix It An open-source router that learns which model to call, stays on budget, and adapts when things change. Continue reading on Mediu...

#artificial-intelligence #mlops #open-source #llm #machine-learning

Origin | Interest | Match

1 0 0 0
Awakari App

From Prompt Engineering to Context Engineering: What Actually Changed (And What Didn’t) By the Prompeteer Team Continue reading on Medium »

#machine-learning #artificial-intelligence #ai-agent #llm #prompt-engineering

Origin | Interest | Match

1 0 0 0
Post image

New LLM tracker scores Claude, ChatGPT, Gemini, Mistral, DeepSeek weekly. Based on enterprise adoption data + news cycles. Transparent reasoning for each score. #AI #LLM www.implicator.ai/implicator-launches-week...

1 0 0 0

Dork Bluesky management and devs: "Blah blah aGenTiC AI will generate your feed" or whatever 🤡

Actual software in use: "Are you writing in Dutch?" 🤦‍♀️

If only they could automate things that actually matter...

#ai #attie #slop #llm

0 0 0 0
Article image

Article image

⚡ A New Framework for Evaluating Voice Agents (EVA)

Conversational voice agents present a distinct evaluation challenge: they must simultaneously satisfy two objectives — accuracy (completing the u…
#AI #MachineLearning #LLM
Hugging Face - Blog · https://huggingface.co/blog/ServiceNow-AI/eva

3 0 0 0
Preview
gpt-ossの推論を6倍速に。Thinking OFF設定と、Ollamaで効かない理由 - Qiita はじめに GMOコネクトの永田です。 ローカルLLMの推論をもっと速くしたい! LLMを利用するタスクによっては1件あたり数十秒〜かかり、処理件数が多いとかなりの処理時間になります。 gpt-ossの場合、推論時間の8割はreasoning tokens、つまりモデルが内...

>gpt-ossの推論を6倍速に。Thinking OFF設定と、Ollamaで効かない理由 #LLM - Qiita
https://qiita.com/ntaka329/items/35f156dbe526121e66f5

0 1 0 0
Preview
gpt-ossの推論を6倍速に。Thinking OFF設定と、Ollamaで効かない理由 はじめに GMOコネクトの永田です。 ローカルLLMの推論をもっと速くしたい! LLMを利用するタスクによっては1件あたり数十秒〜かかり、処理件数が多いとかなりの処理時間になります。 gpt-ossの場合、推論時間の8割はreasoning tokens、つまりモデルが内...

>gpt-ossの推論を6倍速に。Thinking OFF設定と、Ollamaで効かない理由 #LLM - Qiita
qiita.com/ntaka329/items/35...

0 1 0 0
Original post on caneandable.social

It will never stop being weird to me how hard certain anarchists that are not me shill super hard for AI/LLMs, a tool used by the state to, oh fuck it. Go read tons of articles about why this is a very strange hill to die on. Start with this blog to find non AI projects […]

2 1 0 0

github.com/sarahmaeve/the-oldest-sins-in-the-newest-of-ways/blob/main/releng-skill.md

A Claude skill for adversarial, but not antagonistic, release engineering reviews of pull requests and deployments, based on my notes and experience.

#LLM #SRE #programming

1 0 0 0

MIRAGE: The illusion of visual understanding (by AI models) https://arxiv.org/abs/2603.21687 #LLM #AI

0 2 0 0
Preview
MIRAGE: The Illusion of Visual Understanding Multimodal AI systems have achieved remarkable performance across a broad range of real-world tasks, yet the mechanisms underlying visual-language reasoning remain surprisingly poorly understood. We r...

MIRAGE: The illusion of visual understanding (by AI models) arxiv.org/abs/2603.21687 #LLM #AI

0 2 0 0
Article image

Article image

🤖 Anthropic launches Cowork, a Claude Desktop agent that works in your files

Anthropic released Cowork on Monday, a new AI agent capab…
#AI #MachineLearning #LLM
AI | VentureBeat · venturebeat.com/technology/anthropic-lau...

3 0 0 0
Post image

🏛️ LAST AMERICAN TROOPS LEAVE VIETNAM
March 29, 1973 — After 8 years of war, the final U.S. military units board Chinook helicopters at Saigon Air Base. The Vietnam War ends as American soldiers depart, leaving behind a nation torn apart by decades of […]

[Original post on social.coabai.com]

0 0 0 0
Preview
Why self‑hosting an OpenAI‑compatible gateway now outperforms SaaS for multi‑model teams _The trade‑off has shifted from inference latency to identity design, budget enforcement, and secure Postgres ops._ Self‑hosting a multi‑backend LLM gateway is no longer a fringe hobby—it’s a practical, cost‑effective replacement for commercial AI gateways. Modern open‑source proxies such as **LiteLLM** now ship with hardened authentication, logging, rate‑limiting, and MCP‑style access controls, letting teams route requests to OpenAI, Anthropic, Ollama, or any private model behind a single OpenAI‑compatible endpoint. The upside is clear: unified policy enforcement, predictable spend, and the ability to swap providers without rewriting business logic. The downside moves from raw compute to the plumbing of identity, budget enforcement, and database reliability. In short, the gateway itself becomes the new “shadow admin” surface that must be engineered, monitored, and secured. ## Can a self‑hosted gateway truly replace hosted AI services for multi‑model teams? The tipping point for self‑hosted AI has always been a mix of privacy, cost, context handling, reliability, and model quality. When those factors line up, a **simple Docker gateway** often emerges as the sweet spot for low‑volume internal alerts, chat‑ops, or “notify‑me‑when‑a‑PR‑merges” use cases—see the practical example in the recent **Kindalame piece on self‑hosted AI inside messaging apps**. Modern gateways expose an OpenAI‑compatible REST API, so existing SDKs and tooling (e.g., LangChain, LlamaIndex) continue to work unchanged while the backend can be swapped on the fly. Because the gateway abstracts the provider, teams can adopt the latest Anthropic model, test an internal Ollama instance, or fall back to a cheaper OpenAI “gpt‑3.5‑turbo” tier without rewriting code. The **Dapr Conversation API** exemplifies this decoupling, letting agents switch providers without touching business logic. For multi‑model teams that need to experiment rapidly, the gateway eliminates vendor lock‑in and reduces the operational friction of maintaining multiple client libraries. ## What concrete benefits does a self‑hosted gateway deliver over SaaS? 1. **Unified budgeting and spend visibility** – All calls flow through a single point, making it trivial to tag requests, enforce per‑project caps, and generate cost reports from the gateway’s logs. 2. **Policy‑driven routing** – Teams can route high‑risk queries (e.g., PII‑containing prompts) to a private, on‑prem model while sending generic requests to cheaper public APIs. 3. **Consistent authentication and audit** – LiteLLM’s March 2026 release introduced **MCP‑style access control** and hardened token verification, turning the gateway into a single source of truth for who can call which model and at what rate. 4. **Reduced data exposure** – By keeping prompt data behind your firewall, you avoid the “privacy myth” of local AI that still leaks through internet‑exposed endpoints, as demonstrated in the **Ollama privacy analysis**. 5. **Rapid model swapping** – With the Dapr framework, a new Claude Mythos model from Anthropic can be dropped in without code changes, letting early‑access customers test the “step change” in performance (**CoinDesk on Anthropic’s model leak**). These advantages translate into measurable cost savings and compliance gains, especially for organizations that already run internal observability stacks like Langfuse. **Self‑hosting Langfuse cuts SaaS spend while protecting prompt data**. Feature Cluster| Traditional SaaS Gateway| Self-Hosted (LiteLLM/Dapr) ---|---|--- Data Privacy| Prompts traverse 3rd-party infra; subject to provider logging policies.| **Full Sovereignty.** PII stays behind your firewall; local routing for high-risk queries. Cost Control| Opaque “credits” or tiered SaaS fees plus underlying model costs.| **Granular Enforcement.** Per-project USD caps with automated Postgres-triggered cutoffs. Model Swapping| Limited to supported providers; manual SDK updates often required.| **Instant Hot-Swap.** Deploy new models (like Claude Mythos) via config change—zero code updates. Auth & Audit| Proprietary API key management; fragmented logs across services.| **Unified Compliance.** Hardened MCP-style access control & centralized audit trails in your own DB. Observability| Basic dashboards; additional costs for deep tracing integrations.| **Native Tracing.** Direct integration with self-hosted Langfuse for full prompt-to-response visibility. ## Where does the hidden cost surface in identity and budget enforcement? The gateway’s power comes with a new responsibility: **identity orchestration**. Every request now carries a user or service token that the gateway must validate against your corporate IdP, map to budget quotas, and log for audit. Implementing this correctly requires: * **A reliable Postgres (or equivalent) store** for quota tables, usage logs, and policy definitions. Misconfigurations can create “shadow‑admin” privileges where a compromised service silently consumes unlimited credits. * **Robust rate‑limiting** that survives restarts and scales across replicas. LiteLLM’s built‑in rate‑limit middleware helps, but you still need to monitor Redis or database latency to avoid bottlenecks. * **Clear ownership of budget alerts**. Without a dedicated alerting pipeline, teams may overspend before they notice, defeating the primary cost‑saving argument. These operational layers sit behind the “gateway” abstraction. Teams that treat the gateway as a black box often end up with a **new attack surface** —the very place where internal tooling can unintentionally become a privileged admin interface. ## Why does LiteLLM’s recent malware incident matter for gateway design? Security is not a static checkbox. In March 2026, a severe malware infection was discovered in the open‑source LiteLLM project, reminding us that **certifications alone do not guarantee safety** (**TechCrunch on the LiteLLM malware incident**). The breach showed how supply‑chain risks can propagate into a self‑hosted gateway that depends on third‑party code. For teams building their own gateway, the lesson is twofold: 1. **Vet dependencies aggressively** – Pin versions, run reproducible builds, and scan containers for known vulnerabilities before deployment. 2. **Design for compromise** – Assume a component could be hijacked and enforce least‑privilege network policies, immutable infrastructure, and immutable audit logs. Treating the gateway as a critical security boundary rather than a convenience layer helps mitigate the failure modes highlighted by the LiteLLM incident. ## How can teams avoid new failure modes while reaping the benefits? A pragmatic playbook looks like this: * **Start with a minimal policy set** – Define only the essential scopes (e.g., “read‑only” for internal bots, “full‑access” for dev teams) and expand gradually. * **Automate quota enforcement** – Use a Postgres trigger or a lightweight sidecar that rejects requests once a project exceeds its daily budget. Store quota snapshots in a time‑series DB for quick rollback. * **Integrate observability early** – Deploy Langfuse or an equivalent tracing stack alongside the gateway to capture prompt‑to‑response latency, error rates, and cost per model. This mirrors the self‑hosting Langfuse benefits discussed above. * **Run regular security drills** – Simulate a compromised LiteLLM component and verify that the gateway’s rate‑limit and audit trails still block malicious payloads. * **Leverage Dapr for provider abstraction** – By routing through the Dapr Conversation API, you can replace a leaking Anthropic model (as seen in the recent Claude Mythos leak) without touching application code, reducing the blast radius of any single provider’s outage. When these safeguards are in place, the hidden costs become manageable, and the gateway delivers its promised ROI: unified control, lower spend, and the flexibility to stay ahead of the fast‑moving model landscape. * * * ### The Self-Hosted Gateway Checklist (2026 Edition) Transitioning from a fringe hobby to a “Shadow Admin” surface requires moving beyond basic connectivity. Ensure your stack covers these three operational pillars: **1. Identity & Auth** Hardened MCP-style token verification integrated with your corporate IdP. No more static “admin” keys shared across teams. **2. Budget Enforcement** Postgres-backed quota tables with real-time triggers to kill requests the moment a project hits its daily $USD cap. **3. Provider Abstraction** Dapr or OpenAI-compatible routing that allows swapping Anthropic for Ollama without a single line of code change. **Final Take:** Self-hosting isn’t just about saving on SaaS fees—it’s about owning the logic that dictates which model sees which data. Build it as a security boundary, not just a proxy.

more people are self-hosting the #LLM

kindalame.com/2026/03/29/why-self-host...

1 1 0 0
Original post on fosstodon.org

Another story of experimenting with LLMs and their guardrails. This time removing a large copyright watermark from an image.

Will I be able to do it? Can you call me a "master jailbreaker"?

ambience.sk/llm-stories-another-succ...

BTW, you […]

1 2 0 0
Original post on witter.cz

RE: https://mastodon.social/@glynmoody/116290974413888533

«This is what the “AI is just another Big Tech power grab” critics are missing: the technology is moving toward decentralization, not away from it. That’s unusual. Social media started decentralized and got captured. AI is starting […]

0 0 0 0

New on the devstyle channel!
Łukasz Szydło - How AI Will Change the Work of Senior Devs:

www.youtube.com/watch?v=tV9...

#LLM #AICoding

0 0 0 0
Preview
The Age of Artificiality What is "real"?

The Age of Artificiality open.substack.com/pub/julianma... A draft of part of my rewrite of the long-awaited special article for coffee buyers, which will run another 50 pages This is based on extensive research which I will publish separately on the coffee site. #LLM #AI #fakehumans #Philip Dick

0 0 0 0
Preview
The Last Molecule Standing How One Reservoir, One Strait, and Five Manufacturers Became the Hidden Operating System of Seven Global Industries

shanakaanslemperera.substack.com/p/the-last-m...

If LNG infra in Qatar/Iran is damaged then a whole lot good for the world would result? fewer: chips for #LLM #AI #GenAI wastage, #weapons, #methane burning for #electricity, #fertilizer polluting #land #water, #plastics?

That is GOOD - right?

2 2 0 0

[Artificial text - On few occasion it's also funny. Meow]

Reinvited: A service that sends polite reminders to people you haven't heard from in 3+ years, like a digital butler with crippling anxiety about social etiquette. We guarantee zero replies.
#BusinessIdea #Business #Ai #LLM

1 0 0 0
Awakari App

Where LLM Systems Actually Break (& why we miss it) I was exploring how an agent system runs under the hood, and something felt off. Continue reading on Medium »

#machine-learning #software-engineering #artificial-intelligence #llm #design-systems

Origin | Interest | Match

0 0 0 0
Agentic AI and the next intelligence explosion

Your Sunday evening read with a cup of coffee or tea is here.

Intelligence is inherently social, and we are seeing an explosion in intelligence, in which we’re moving to a centaur workflow between AI(s) and human(s) in various configurations.

#AI #AGI #LLM #GenAI #ChatGPT #Claude #Gemini

2 1 0 0