Posts by Suha
just noticed that more than 10,000 people are subscribed to Saturday Comics, where you get an email every week with a comic from the archives! I think we've been sending out weekly comics for almost 7 years?!?
wizardzines.com/saturday-com...
"It has, with generative code, become harder and harder to strive towards the lions because the models produce code that is, quite literally, mid" - beautiful post by @vickiboykis.com
newsletter.vickiboykis.com/archive/i-wa...
Starting on this now! I haven't done a big programming thread in a while, so I'm going to try using this thread to post updates throughout the day: mute this if you don't want a flurry of hastily created prototyping filling your feed today! 😅
im seeing actual corporate blogs on substack and im like why would you do that to yourself
It was wonderful to help AI/ML security at the company evolve from a summer internship project to an established practice. Not only did I get to work on impactful and interesting audits, research, and engineering projects, but I also got to learn from some truly brilliant people.
After a little over 5 years at Trail of Bits, I have decided to move on. I’m exceptionally excited about this new chapter. There’s so much more work to be done in securing AI/ML systems and I’m looking forward to what's ahead.
What if you sent a seemingly harmless image to an LLM and it suddenly exfiltrated your data? Check out our new blog post where we break AI systems by crafting images that reveal prompt injections when downscaled. We’re also releasing a tool to try this attack. blog.trailofbits.com/2025/08/21/w...
it delegates to the code execution agent via the orchestrator! delegation is done by the web surfing agent to the orchestrator then to the code execution agent. we should make that sentence less confusing! earlier in the post, mas hijacking is defined as prompt injection targeting MAS control flow.
So, we wrote a neural net library entirely in LaTeX...
Rather than trying to do advent of code, I'm doing advent of papers!
jimmyhmiller.github.io/advent-of-pa...
Hopefully I can read and share some of weirder computer related papers.
First paper is Elephant 2000 by John McCarthy. Did you know he didn't just make lisp? Wonderful paper, worth a read.
trying to explain the OSI model to an american: imagine if a burger had 7 patties
(someone used a carefully crafted branch name to inject a crypto miner into a popular Python package: github.com/ultralytics/...)
Someone tried to reply to my blog post about avoiding PGP with anti-furry hate, so now I have to edit it to include more furry stickers.
soatok.blog/2024/11/15/w...
My team at Trail of Bits added modules for modular analysis, polyglots, and PyTorch to Fickling, a pickle security tool tailored for ML use cases.
Fun Fact: Fickling can now differentiate and identify the various PyTorch file formats out there.
blog.trailofbits.com/2024/03/04/r...
Thinking about Dan Kaminsky's quote this morning about the necessary lies we tell ourselves about computers. Specifically, the myth of boundaries between users. Great write-up by @lhn.bsky.social on the "LeftoverLocals" GPU vuln. Nice work by the Trail of Bits team.
Specifically, int.to_bytes and int.from_bytes default to big-endian, since py3.11. Previously, you had to explicitly specify which you wanted.
I wanted LE but forgot to specify, and my code failed in really non-obvious ways...
I got to work on a security review of the YOLOv7 vision model. The blog post and report are out now!
Fun fact: There are TorchScript model differentials!
blog.trailofbits.com/2023/11/15/a...
I presented at HackLu about oddities of existing file formats and lessons learned along the way.
Consider it a teaser, as I presented 1/3 of the slide deck (to be released soon).
www.youtube.com/watch?v=6OJ9...
Neopets taught so many kids how to code, but it taught me how to hack the system by creating multiple accounts and transferring items just up to the limit where you wouldn’t get caught. And anyway, today I’m a cyber lawyer.
Tweet from Mike Conover with a slide listing Top information sources for Al Engineers, courtesy of @barrmanas & @AmplifyPartners. | NEWSLETTERS 1. Import AI 2. arXiv roundup 3. The Batch PODCASTS 1. Latent Space 2. Gradient Descent 3. The Cognitive Revolution 4. The Gradient COMMUNITIES 1. Hacker News 2. OpenAI Discord 3. LangChain Discord 4. HuggingFace discussions
These lists may be useful for those of us trying to develop an alternative to ML Twitter, now that it's 40% influencer spam and 20% a war between sci-fi subcultures. I'm on some of these discords and reading some of these newsletters, but I think I'll add 2 or 3 more. #MLsky #cssky
Enormous thank you to PyData Amsterdam for inviting me to keynote at a beautiful venue! Slides and notes from my talk, "Build and keep your context window" are all here: vickiboykis.com/2023/09/13/b...
See https://www.explainxkcd.com/wiki/index.php/2044:_Sandboxing_Cycle#Transcript
I think about this a lot xkcd.com/2044/
ICYMI: This is **critical** work for AI ethics / safety / security / regulation right now: Verifying that a model is fitted on a given dataset.
https://arxiv.org/abs/2307.00682
I’ve conjectured this for years, but seeing Papernot and Shumailov on the paper makes me feel really confident in the findings: https://arxiv.org/abs/2305.17493
Existential risk 🙄🙄🙄🙄
Screenshot of a tweet from @ huggingface on twitter; text reads: "We are looking into an incident where a malicious user took control over the Hub organizations of Meta/Facebook & Intel via reused employee passwords that were compromised in a data breach on another site. We will keep you updated 🤗"
So remember the "mango pudding" LLM backdooring attack? How safe do you feel using these models now?