Streaming Reinforcement Learning (RL) is a huge challenge: transitions are used once and discarded immediately. This makes agents extremely sample-inefficient. But what if we could "squeeze" more information out of every single frame?
Check out our latest paper!
Posts by Saurav Jha
New work, just accepted @ICLR: "The Expressive Limits of Diagonal SSMs for State-Tracking"
We give a complete characterization of what diagonal SSMs can and cannot compute on state-tracking tasks and the answer is deeply connected to group theory.
π§΅π
Can LLMs play Hangman? Spoiler alert: Not yet.
Check out βLLMs Canβt Play Hangman: On the Necessity of a Private Working Memory for Language Agentsβ, led by Davide Baldelli, Ali Parviz, AmalZouaq and Sarath Chandar.
Can LLMs become CAD designers?
Check out βCADmium: Fine-Tuning Code Language Models for Text-Driven Sequential CAD Designβ, which is now published in Transactions on Machine Learning Research (TMLR)!
Life update - last month I moved to #montreal π¨π¦ from #Sydney π¦πΊ to kick off my @ivado.bsky.social postdoc fellowship at @mila-quebec.bsky.social. Must say I am constantly amused by: 1. How walkable the city is.
2. How easy is it to reach out to diverse research communities within #mila ! π
π Happy to share that our paper βMining your own secrets: Diffusion Classifier scores for Continual Personalization of Text-to-Image Diffusion Modelsβ has been accepted to #ICLR2025!
π The work results from my #Sony summer internship in the stunning #TokyoπΌ city
Preprint: arxiv.org/pdf/2410.00700
I ran across a busy Sander at a #neurips party with a similar question - he was still patient enough to explain stuff. This talk further clarifies a good amount of my doubts. Recommend watching if you're working on diffusion / LLMs for generation!
I validate this