Advertisement · 728 × 90

Posts by Michael Hind

Every Eval Ever | EvalEval Coalition

🚀 Launching Every Eval Ever: Toward a Common Language for AI Eval Reporting 🚀

A shared schema + crowdsourced repository so we can finally compare evals across frameworks and stop rerunning everything from scratch 🔧

A tale of broken AI evals 🧵👇

evalevalai.com/projects/eve...

2 months ago 12 4 1 4

I'm lucky to be a part of this wonderful collaboration to improve the transparency and use of AI benchmarks. research.ibm.com/blog/documen...

4 months ago 0 0 0 0
Preview
How IBM’s Kush Varshney became an iconic ’test’ photo The IBM Fellow reflects on copyright law, generative AI, and how he became the face of the modern camera man

An interesting backstory of a common test photo sparked another photo (of @krvarshney.bsky.social) in another dataset. research.ibm.com/blog/kush-va...

9 months ago 2 0 0 0
Preview
University of Notre Dame and IBM Research build tools for AI governance - Lucy Family Institute for Data & Society Expanding into virtually all aspects of modern society, AI systems are transforming everything from education to healthcare, but how trustworthy are the vast

I'm excited to be a part of this great collaboration with colleagues at IBM Research and Notre Dame. lucyinstitute.nd.edu/news-events/...

9 months ago 2 0 0 0
Preview
IBM enhances the capabilities of watsonx.governance with the new Model Risk Evaluation Engine We're excited to announce the Model Risk Evaluation Engine, a new tool in watsonx.governance that can measure measures risks of foundation models by computing metrics related to risk dimensions from t...

Are you wondering how you can evaluate some of the risks of a foundation model before you deploy it? Read on .... www.ibm.com/new/announce...

1 year ago 3 0 0 0
DeepSeek-V3-0324, Gemini Canvas and GPT-4o image generation
DeepSeek-V3-0324, Gemini Canvas and GPT-4o image generation YouTube video by IBM Technology

I'm on the IBM Mixture of Experts podcast wearing a safety vest. We talk about all the new things in AI this week. I also connect to older work by IBM Fellows Irene Greif, Bob Dennard, Rolf Landauer, and Charlie Bennett and to Mauro Martino's new AI-generated film. www.youtube.com/watch?v=CgqH...

1 year ago 2 2 0 0
Post image

Happy to see Granite Guardian models atop the GuardBench leaderboard, including in non-English languages.

This benchmark was just released. Read about it here: www.linkedin.com/posts/eliasb....

1 year ago 3 1 0 0
Preview
Decolonial AI Alignment by Kush Varshney (IBM Research, US)

A summary of decolonial AI alignment in the Human-Centered AI publication on Medium. Thanks to @jweisz3.bsky.social for asking me to write it, and for editing the piece. medium.com/human-center...

1 year ago 5 2 0 0
Preview
Work for me, but don’t pretend to be me: on transparency of identity in AI… | Thomas Hampp You may trust your concierge with the keys to your house, but would you allow them to sign with your name? What about giving your AI agent the API keys to your accounts? In my article, I discuss an i...

I'm happy to see my former IBM colleague raise this important issue regarding Agentic systems. www.linkedin.com/posts/thomas...

1 year ago 0 0 0 0
Preview
Erik Miehling on LinkedIn: AI development is currently overly focused on individual model… AI development is currently overly focused on individual model capabilities, often ignoring broader emergent behavior, leading to a significant underestimation…

From Erik Miehling (www.linkedin.com/posts/erik-m...)

"AI development is currently overly focused on individual model capabilities, often ignoring broader emergent behavior, leading to a significant underestimation of the true capabilities and associated risks of agentic AI."

1 year ago 6 2 0 0
Advertisement
Post image

Four exciting things to share about watsonx.governance and Granite Guardian. Fun times in AI safety! See thread for the details.

1 year ago 1 1 1 0

"... We'd love your feedback! Try the code, explore the Hugging Face space, and join us in building a stronger governance framework for AI."

www.linkedin.com/posts/elizab...

1 year ago 0 0 0 0

From Elizabeth Daly: "This week we are releasing, Risk Atlas Nexus, github.com/IBM/risk-atl..., an open source project that provides tooling to help bring together disparate resources related to governance of foundation models. ... "

1 year ago 0 0 1 0
Preview
Bringing reasoning to Granite We’re excited to announce a preview release of new reasoning capabilities in our Granite family of large language models.

"While techniques such as the ones used by R1 can degrade model safety, our preview release shows that reasoning and safety don’t have to be a trade-off."
www.ibm.com/new/announce...

1 year ago 4 2 0 0
Preview
Transparency in Discussion: Improving Transparency and Accountability in AI Implementations Podcast Episode · Humanitarian AI Today · 11/24/2024 · 1h 19m

It was a pleasure to join the panel discussion on Humanitarian AI Today podcast below, moderated by Brent Phillips: podcasts.apple.com/us/podcast/t...

1 year ago 2 1 0 0
Preview
The Key to How IBM's Granite 3.1 is Advancing Enterprise AI IBM’s new Granite 3.1 addresses key enterprise needs, including expanded context handling, multilingual support, new tools and AI agent development

"IBM has equipped the Granite Guardian 3.1 models with the ability to detect hallucinations in AI agent workflows. This feature provides oversight of an AI agent completing a task, monitoring for fabricated information or incorrect function calls." technologymagazine.com/articles/the...

1 year ago 6 2 0 0
ACM FAccT - 2025 CFP

Reminder: The #FAccT2025 submission deadlines are roughly one month away! Abstracts are due January 15th and full papers on January 22nd. See the full CfP here: facctconference.org/2025/cfp

1 year ago 26 12 0 0
Preview
What's Simmering? (Part 1) By The Human-Centered Trustworthy AI Teams at IBM Research Earlier in the year, I posted a non-exhaustive narrative of what's simmering in Human-Centered Trustworthy AI at IBM Research. I'm at it again.

www.linkedin.com/pulse/whats-...

1 year ago 1 0 0 0
Preview
Granite Guardian Demo - a Hugging Face Space by ibm-granite demo

I showed this cool demo last week @neuripsconf.bsky.social Now we have a public version on Hugging Face that you can play with to see the "judge" model in action. huggingface.co/spaces/ibm-g...
Enjoy!

Open source repo & benchmarks: github.com/ibm-granite/...

1 year ago 7 4 0 0
Post image

Now posted at the under construction booth 😀 our demo lineup for Tuesday. Looking forward connecting with you at the IBM booth @neuripsconf.bsky.social

1 year ago 8 2 0 0
Advertisement
IBM Researchers setting up a booth at a convention center. Some are wearing safety gear and some are not.

IBM Researchers setting up a booth at a convention center. Some are wearing safety gear and some are not.

It is @neuripsconf.bsky.social booth setup day! Among Ambrish Rawat, @bhoov.bsky.social, and @wernergeyer.bsky.social, who do you think is *not* an author of the Granite Guardian technical report we released today? (Hint: Granite Guardian helps make any LLM safer.)

Link: github.com/ibm-granite/...

1 year ago 6 2 0 0
Tips

If you’re headed to NeurIPS 2024, and want to learn about IBM Research Human-Centered Trustworthy AI, there are many many opportunities to do so.

1. Start with the official NeurIPS explorer by @henstr.bsky.social and @benhoover.bsky.social. It is infoviz par excellence. neurips2024.vizhub.ai

1 year ago 9 3 1 0

What are the desirable properties of AI metrics for such tests? What about summarizing these metrics for non-technical stakeholders?

1 year ago 1 0 0 0

... or when a physician tries to diagnose the health of a new patient by performing various diagnostic medical tests (blood tests, x-rays, etc).
What happens when one applies these ideas to AI models? How can it be helpful? How can it be misleading? What role could this play in regulations?

1 year ago 1 0 1 0

The work explores the challenges of testing for AI risks without have any information of how the model was developed, such as when one purchases a model from a 3rd party or open source. Similar to how a home inspector is asked to inspect a home without knowing its construction history.

1 year ago 1 0 1 0
Preview
Quantitative AI Risk Assessments: Opportunities and Challenges Although AI systems are increasingly being leveraged to provide value to organizations, individuals, and society, significant attendant risks have been identified and have manifested. These risks have...

I'm happy to announce a significant revision of our paper describing opportunities and challenges of quantitative AI risk assessments, also known as automated red-teaming: arxiv.org/abs/2209.06317

1 year ago 6 3 1 0
Overview of paper browser. A cluster for reinforcement learning is selected.

Overview of paper browser. A cluster for reinforcement learning is selected.

Paper Browser: only papers assigned to "physical models - physics" are shown.

Paper Browser: only papers assigned to "physical models - physics" are shown.

Paper Browser: Filtered by author "Hoover" and detail is shown

Paper Browser: Filtered by author "Hoover" and detail is shown

Paper Brower: ZOOOOM in

Paper Brower: ZOOOOM in

🎺 Here comes the official 2024 NeurIPS paper browser:
- browse all NeurIPS papers in a visual way
- select clusters of interest and get cluster summary
- ZOOOOM in
- filter by human assigned keywords
- filter by substring (authors, titles)

neurips2024.vizhub.ai

#neurips by IBM Research Cambridge

1 year ago 60 22 5 4

I enjoyed my recent interview on the AI Risk Reward podcast with host Alec Crawford.
You can hear it here: podcasts.apple.com/us/podcast/t...

2 years ago 3 0 0 0