Advertisement · 728 × 90

Posts by Kathy Baxter

Post image Post image

When LLM agents talk to each other, a new failure emerges: agents abandon their assigned roles and mimic their conversational partner instead. Rates reach as high as 70% with major models — and increased reasoning doesn’t help.

🧵 arxiv.org/abs/2511.09710

1 month ago 23 12 1 4

Our work on "Echoing" was accepted at the Agents in the Wild workshop at ICLR 2026

When LLM agents interact without oversight, they can abandon their assigned roles — and standard metrics won't catch it.

#ICLR2026 #FutureOfAI #EnterpriseAI

1 month ago 3 1 0 0
Preview
Poisoning the Well: Search Agents Get Tricked by Maliciously Hosted Content Is your AI agent being tricked? Our new research explores "well poisoning" attacks and why 80% of agents fail to detect malicious content. Learn more about Salesforce’s security-first approach.

Poisoning the Well: Search Agents Get Tricked by Maliciously Hosted Content sforce.co/40Z1Gi5

AI search agents can be fooled by planted fake content, ~80% of queries returned the attacker's answer in controlled tests.

Authors: Shafiq Joty, Xuan Phi Nguyen, Shrey Pandit, Yifei Ming

1 month ago 4 1 0 0
Preview
MAS-ProVe: Understanding the Process Verification of Multi-Agent Systems Multi-Agent Systems (MAS) built on Large Language Models (LLMs) often exhibit high variance in their reasoning trajectories. Process verification, which evaluates intermediate steps in trajectories, h...

(🧵 1/5) MAS-ProVe — the first systematic study of process verification for multi-agent systems.

Paper: arxiv.org/abs/2602.03053

Do multi-agent systems actually benefit from automatic process-level verification? This paper puts that question to a rigorous test.

1 month ago 4 1 1 0
InterviewSim: A Scalable Framework for Interview-Grounded Personality Simulation Simulating real personalities with large language models requires grounding generation in authentic personal data. Existing evaluation approaches rely on demographic surveys, personality questionnaire...

(🧵 1/5) InterviewSim: A Scalable Framework for Interview-Grounded Personality Simulation bit.ly/3P4cQj9

🔍 Can LLMs faithfully simulate real people? Most evaluations never check against what individuals actually said.

1 month ago 3 1 1 0
Preview
Product Management Director/Sr. Director - Responsible AI (RAI) Director or Sr. Director - Product Management, Responsible AI (RAI) - Technical We are seeking an experienced, technically fluent Director / Sr. Director of Product Management to deliver Responsible A...

@salesforce.com #EthicalUse team is hiring a Dir./Sr. Dir. of #ResponsibleAI!
The ideal candidate has:
- PM Leadership: Proven industry experience with AI/ML platforms, GAI, agents
- RAI Expertise: Deep familiarity with #RAI concepts, governance, & assurance
careers.salesforce.com/en/jobs/jr32...

3 months ago 1 1 0 0
Preview
Digital surveys may have hit the AI point of no return A new study shows bots can be indistinguishable from humans in survey responses.

New Article Alert By me @fastcompany.com 🚨
"Digital surveys may have hit the #AI point of no return."
Quantitative surveys can now be taken by AI that undetectably mimics humans. This can compromise results that will impact decisions based on these outcomes. A way forward? Qualitative Research!

3 months ago 15 8 0 0
The AI Dilemma — with Tristan Harris
The AI Dilemma — with Tristan Harris YouTube video by The Prof G Pod – Scott Galloway

1/ Interesting @profgalloway.com conversation with Tristan Harris on AI risks. From companion bots over job losses to AGI. I don’t agree with all of Tristan Harris’ views, but he excels at painting a very clear and simple picture of the risks we face.

www.youtube.com/watch?v=MLvx...

4 months ago 1 1 1 0
Preview
Washington’s AI Task Force delivers policy recommendations to promote innovation and protect individual rights | Washington State The Washington State Artificial Intelligence (AI) Task Force released its interim report on Monday, which delivers eight key policy recommendations to the governor and the legislature aimed at…

Washington’s AI Task Force has issued interim recommendations including NIST-aligned principles, dataset transparency rules, high-risk governance requirements, and new AI reporting standards. Source: www.atg.wa.gov/news/news-re... #AIPolicy

4 months ago 1 1 0 0
Preview
OpenAI violated copyright law, German court rules A German court has ruled against OpenAI in a landmark copyright case.

A Munich court ruled that OpenAI violated German copyright law in a case brought by GEMA over training on protected song lyrics, ordering the company to pay damages. mashable.com/article/open... #AIPolicy

4 months ago 3 1 0 0
Advertisement

what really gets me every damn time is why "AI" products built with children in mind are rarely vetted or regulated. even simple household appliances go through so much testing, assurance, & regulatory process

and now the industry is lobbying so hard, we've regressed back to a decade

4 months ago 59 15 1 1
Preview
Montreal AI Ethics Institute | Democratizing AI ethics literacy

Proud to be a contributor to @mtlaiethics.bsky.social's State of #AI #Ethics Report Vol. 7! 58 global contributors, 17 topics, & 48 essays. It features grounded examples, community-driven solutions, & insights from across sectors. 🌍
#responsibleAI #AIgovernance #agenticAI
👉Read: montrealethics.ai

5 months ago 6 0 0 0
Video

🎉Today, we released a series of five Responsible AI courses that you can take for free: alltechishuman.org/rai-courses

Principles of Responsible AI

The History of Responsible AI

Operationalizing AI Governance (2)

Governing Agentic AI Systems

5 months ago 10 5 0 0
Video

Our latest Responsible Tech Guide is out! We cover the state of Responsible Tech, pathways and pipelines, and the intersections between Responsible AI, Trust & Safety, and Public Interest Technology.

DOWNLOAD: alltechishuman.org/responsible-...

6 months ago 9 3 0 0
Video

I'm excited about #NVIDIA #GTC DC this week! Please join me, Elham Tabassi (@brookings.edu), Chris Meserole (Frontier Model Forum), and @roar817.bsky.social (NVIDIA) for our panel on "Foundational Concepts in #AISafety on Wed., Oct. 29th at 3p ET!
www.nvidia.com/gtc/dc/sessi...
#AI #ResponsibleAI

5 months ago 5 0 0 0
32 people standing and sitting in a ballroom in front of a screen with the World Economic Forum logo and the title "Faith in Action Convening: Embedding Dignity in the Intelligent Age."

32 people standing and sitting in a ballroom in front of a screen with the World Economic Forum logo and the title "Faith in Action Convening: Embedding Dignity in the Intelligent Age."

At the @weforum.org Global Futures Council Summit last week, I was honored to speak at the "Faith in Action: Embedding Dignity in the Intelligent Age" dinner. The event convened global leaders working at the intersection of #AI & moral insight to explore ways to embed human dignity in AI development

6 months ago 4 1 0 0
Preview
Salesforce earns first ISO 42001 certification Salesforce earns ISO 42001 certification for Agentforce, Einstein, and Slack AI, proving its commitment to ethical, secure, and transparent AI.

I'm thrilled to announce that @salesforce.com has obtained the ISO/IEC 42001 certification for our #AI Management System (AIMS), the first internationally recognized standard for #ResponsibleAI governance! 🎉 www.salesforce.com/blog/salesfo...

6 months ago 3 0 0 2
Video

Pivot is going on tour!

Catch us in: Toronto, Boston, NYC, DC, Chicago, San Francisco and LA.

Kara Swisher. Scott Galloway. 7 Cities. 7 Days. Endless jokes and bickering.

Go to PivotTour.com for tickets!

6 months ago 180 17 31 5
Video

Big news! I've been working with @sladner.bsky.social on "Off the Path" -- a limited series podcast where we talk about user research careers, especially our own. Episodes come out October 10. In this hype video, I reflect on why I decided to take on this new project together.

6 months ago 11 5 1 1
Preview
Agents of Trust | Credo AI Trust Summit 2025 Credo AI Trust Summit 2025. Where intelligent systems and human leadership align to make trust the foundation of progress.

Excited to attend the Credo AI #AI #Trust Summit tomorrow (Tuesday)! Who else will be there?

www.credo.ai/apply-summit...

6 months ago 1 0 0 0
Advertisement

That was a fantastic event and kudos to the entire @axios.com reporting staff for asking the tough questions!

7 months ago 0 0 0 0
Preview
WATCH: The Axios AI+ DC Summit featuring Anthropic's co-founders, key senators and more Tune in to hear from top voices from government, tech and business exploring the urgent questions around AI regulation, geopolitics, misinformation and the real world impact of AI live from Axios' AI+...

I'm excited to dial into the live stream of the @axios.com #AI + DC Summit with @inafried.bsky.social at 11a PT / 2p ET!
www.axios.com/2025/09/10/w...

7 months ago 5 1 1 1
Preview
Meta suppressed research on child safety, employees say The company’s lawyers intervened to shape research that might have shed light on risks in virtual reality, four current and former staffers have told Congress. Meta denies the allegations.

A German teen told Meta's child safety researchers his under-10 little brother had been sexually propositioned multiple times on its VR platform.

Meta deleted the evidence.

New internal whistleblower docs indicate that was part of a broader cover-up: www.washingtonpost.com/investigatio...

7 months ago 3098 1551 81 223

@weforum.org GFC Report "Synthetic Data: The New Data Frontier." Bias, scarcity, & privacy constraints limit the potential of real-world data. Synthetic data offers a path forward. Learn more! @digitalarun.bsky.social @sarahooker.bsky.social @nighatdad.bsky.social
reports.weforum.org/docs/WEF_Syn...

7 months ago 2 1 0 0

AI companies should change the way their product works in a fundamental way. It should engage like a computer not a human, they don’t have minds, can’t think. They should work and sound like a computer. Prevent tragedy like this.

7 months ago 29 7 4 3
Preview
How to stop AI from seeming conscious | IBM What happens when people mistake an AI bot for a conscious being? Researchers say the consequences can be dangerous.

Personas & long-term memory increase the risk of people connecting emotionally with #AI & believing it is conscious. Those features may come with benefits but they also pose unintended risk to ppl's wellbeing. The "things" we risk breaking when we move fast, are PEOPLE.
www.ibm.com/think/news/h...

7 months ago 3 0 1 0
Preview
OpenAI Says It's Scanning Users' ChatGPT Conversations and Reporting Content to the Police OpenAI has authorized itself to call law enforcement if users say threatening enough things when talking to ChatGPT.

futurism.com/openai-scann...

7 months ago 7 2 0 0

I’m paying you a million dollars per year. Just solve ASI already. What’s the problem?

7 months ago 35 5 3 1

Amen to this. I wrote about this problem in "Artificial Unintelligence" in 2018; it's only gotten worse in the AI era.

7 months ago 107 24 1 1
Advertisement