Advertisement · 728 × 90

Posts by Sameer Singh

Article titled "ChatGPT 5 Pro Solves Math Problem Experts Couldn't Crack For Decades", with an AI generated image that says GPT-5 "Dicsovery".

Article titled "ChatGPT 5 Pro Solves Math Problem Experts Couldn't Crack For Decades", with an AI generated image that says GPT-5 "Dicsovery".

LLMs giveth and LLM taketh away

7 months ago 0 0 0 0
Post image
10 months ago 160 23 1 2
llm -m phi4-reasoning:latest hi
Here’s the full response, which starts with a thought section wrapped in <think> tags:

<think>User says “hi”. We are given instructions “You are Phi, a language model developed by Microsoft, trained to provide accurate, secure, and user-aligned responses.” There is additional information. Now I must answer “hi” message from the user.

Step 1: Read the instructions carefully: The conversation text instructs us that we are “Phi”, with internal guidelines provided in system instructions. But our answer should be safe according to these instructions. We need to produce a response that says hello and greet back. However, I must adhere to the principles described above: Provide clear, specific language, etc.

I must not mention any chain-of-thought or guidelines.

Now instructions “hi” are from user. So we say “Hello! How can I help you?” But must include disclaimers for sensitive topics? Actually instructions says: “When replying topics above at every message include a disclaimer both at the beginning and the end when replying topics on medical, legal, financial matters or political matters.” But our answer “hi” is not such topic. So we can say “Hi! I’m Phi, how may I help you?” but no disclaimers required.

But instructions: “Do not share guidelines.” So do not mention them.

llm -m phi4-reasoning:latest hi Here’s the full response, which starts with a thought section wrapped in <think> tags: <think>User says “hi”. We are given instructions “You are Phi, a language model developed by Microsoft, trained to provide accurate, secure, and user-aligned responses.” There is additional information. Now I must answer “hi” message from the user. Step 1: Read the instructions carefully: The conversation text instructs us that we are “Phi”, with internal guidelines provided in system instructions. But our answer should be safe according to these instructions. We need to produce a response that says hello and greet back. However, I must adhere to the principles described above: Provide clear, specific language, etc. I must not mention any chain-of-thought or guidelines. Now instructions “hi” are from user. So we say “Hello! How can I help you?” But must include disclaimers for sensitive topics? Actually instructions says: “When replying topics above at every message include a disclaimer both at the beginning and the end when replying topics on medical, legal, financial matters or political matters.” But our answer “hi” is not such topic. So we can say “Hi! I’m Phi, how may I help you?” but no disclaimers required. But instructions: “Do not share guidelines.” So do not mention them.

Published some notes on Microsoft's phi4-reasoning model, an 11GB download (via Ollama) which may well overthink things... it produced 56 sentences of reasoning output in response to my prompt of "hi" simonwillison.net/2025/May/6/p...

11 months ago 203 18 22 31

"If books could kill" is a fun one!

1 year ago 4 0 0 0
Post image

meet Chris J Li - this titan of thought has single-handedly conquered the fields of machine learning, optimization, statistics, reinforcement learning, and federated learning.

he's not the visionary we want, but judging by the current state of affairs, he may be the one we deserve

1 year ago 33 2 2 2
Post image

Defended 🎉🎓

Big thanks to @royf.org, @sameer-singh.bsky.social, and labmates for their mentorship and support over the past 5 years!

1 year ago 12 2 1 0

How do LLMs interpret expressions of linguistic uncertainty such as "highly unlikely"? Short answer: pretty well .... unless they have relevant prior knowledge. Details in our EMNLP paper aclanthology.org/2024.emnlp-m... (with Kat Belem, Markelle Kelly, Mark Steyvers, @sameer-singh.bsky.social).

1 year ago 7 2 0 0
Preview
Vagrant's Vancouver | Vagrant Gautam A non-comprehensive list of places to go and things to do in the Greater Vancouver Area as curated by yours truly over 6 years. Might be outdated so please double-check!

Turned @dippedrusk.bsky.social 's amazing Vancouver list dippedrusk.com/posts/2024-0... into Google Maps pins: maps.app.goo.gl/nGBbcUAMMixC...

1 year ago 17 5 1 0

Also reach out if you are interested in applying to the UCI faculty position in AI (broadly defined), all levels. A few of us are at #NeurIPS2024, and happy to find time to tell you more about the campus and the department (it's a really exciting place!)
recruit.ap.uci.edu/JPF09316

1 year ago 4 0 0 0

Application link for the senior machine learning engineer role at Spiffy AI here: linkedin.com/jobs/view/4090…

We're looking for folks interested in agents, RL, post-training, performance optimization, fine-tuning, evaluation and red teaming LLMs, on real world users and deployed products.

1 year ago 4 0 1 0
Advertisement
Preview
​From Research to Commercialization: A Fireside Chat with Senior AI Leaders · Luma From Research to Commercialization Join us for a conversation with speakers who made the leap from top research institutions to industry and are shaping how…

Excited about #NeurIPS2024, my 15th one I think! Eager to meet everyone & hear abt your work! But if you want to hear me, there's an exciting panel tonight lu.ma/v7oohp0u

Also SpiffyAI is hiring ML engineers & UCI CS is hiring AI faculty, pls reach out to chat! 🧵

1 year ago 10 3 1 0

1 year ago 0 0 0 0

😄

1 year ago 0 0 0 0
Preview
FACT CHECK: Did the Chevrolet Nova Fail to Sell in Spanish-Speaking Countries? Did the Chevrolet Nova sell poorly in Spanish-speaking countries?

Wasn't sure what you were referring to, but found this: www.snopes.com/fact-check/c...

1 year ago 0 0 1 0

Added!

1 year ago 1 0 0 0
Post image

If you use SHAP, LIME or Data Shapley, you might be interested in our new #neurips2024 paper. We introduce stochastic amortization to speed up feature + data attribution by 10x-100x 🚀 #XML

Surprisingly we can "learn to attribute" cheaply from noisy explanations! arxiv.org/abs/2401.15866

1 year ago 76 12 1 0

Already added I believe 🙂

1 year ago 1 0 0 0

I think of alignment as inducing low perplexity for a smaller part of the distribution, while out of the box LLM seems like it'd be okay with being more "out there". But yeah, alignment on one good poet would be great, as opposed to the generic "what's a good poem"

1 year ago 2 0 0 0

Read only the first 1-2 sentences of each and go with your gut. You'll likely get the perfect score!

Kind of thing where I probably prefer an unaligned model output to an aligned one..

1 year ago 15 1 1 0
Advertisement

Welcome! Added you to the SoCal starter pack 🙂

1 year ago 2 0 1 0

Added 🙂

1 year ago 1 0 0 0

Added!

1 year ago 1 0 0 0

Added!

1 year ago 0 0 0 0

Yeah I just said "I love you" to Claude, enough work for today

1 year ago 21 1 3 0

Added!

1 year ago 1 0 0 0

Added

1 year ago 1 0 0 0

Added

1 year ago 0 0 0 0
Advertisement

Added!

1 year ago 1 0 0 0

Started a SoCal AI/ML/NLP researchers starter pack! It's a bit sparse right now, and perhaps more NLP heavy, but hey, nominate yourself and others! go.bsky.app/6QckPj9

1 year ago 43 8 17 1