Which, whose, and how much knowledge do LLMs represent?
I'm excited to share our preprint answering these questions:
"Epistemic Diversity and Knowledge Collapse in Large Language Models"
๐Paper: arxiv.org/pdf/2510.04226
๐ปCode: github.com/dwright37/ll...
1/10
Posts by Srishti
Happy to share that our work on multi-modal framing analysis of news was accepted to #EMNLP2025!
Understanding news output and embedded biases is especially important in today's environment and it's imperative to take a holistic look at it.
Looking forward to presenting it in Suzhou!
๐ Looking for PhD opportunities in #NLProc for a start in Spring 2026?
๐๏ธ Add your expression of interest to join @copenlu.bsky.social here by 20 July: forms.office.com/e/HZSmgR9nXB
Selected candidates will be invited to submit a DARA fellowship application with me: daracademy.dk/fellowship/f...
๐ฃ I am happy to support Ph.D applications to the Danish Advanced Research Academy. My main areas of research include multimodal learning and tokenization-free language processing. Feel free to reach out if you have similar interests! Applications due August 29 www.daracademy.dk/fellowship/f...
Congratulations Andrew Rabinovich (PhD โ08) on winning the Longuet-Higgins Prize at #CVPR2025! (1/2)
My favorite part of going to conferences: @belongielab.org alumni get-togethers! A big thank you to Menglin for coordinating the lunch at @cvprconference.bsky.social ๐
Left: Tsung-Yi Lin, Guandao Yang, Katie Luo, Boyi Li; Right: Menglin Jia, Subarna Tripathi, Ph.D., Srishti, Xun Huang
๐ Technical practitioners & grads โ join to build an LLM evaluation hub!
Infra Goals:
๐ง Share evaluation outputs & params
๐ Query results across experiments
Perfect for ๐งฐ hands-on folks ready to build tools the whole community can use
Join the EvalEval Coalition here ๐
forms.gle/6fEmrqJkxidy...
Please join us for the FGVC workshop at CVPR 2025 @cvprconference.bsky.social on Wed 11th of June. The full schedule and list of fantastic speakers can be found on our website:
sites.google.com/view/fgvc12
Can you train a performant language model using only openly licensed text?
We are thrilled to announce the Common Pile v0.1, an 8TB dataset of openly licensed and public domain text. We train 7B models for 1T and 2T tokens and match the performance similar models like LLaMA 1 & 2
"Large [language] models should not be viewed primarily as intelligent agents but as a new kind of cultural and social technology, allowing humans to take advantage of information other humans have accumulated." henryfarrell.net/wp-content/u...
Would you present your next NeurIPS paper in Europe instead of traveling to San Diego (US) if this was an option? Sรธren Hauberg (DTU) and I would love to hear the answer through this poll: (1/6)
"I donโt want to just be entering text prompts for the rest of my life."
I spoke to political cartoonists, including Pulitzer-winner Mark Fiore, about how they are using AI image generators in their work. My latest for @niemanlab.org.
www.niemanlab.org/2025/06/i-do...
Culture is not trivia: sociocultural theory for cultural NLP. By Naitian Zhou and David Bamman from the Berkeley School of Information and Isaac L. Bleaman from Berkeley Linguistics.
There's been a lot of work on "culture" in NLP, but not much agreement on what it is.
A position paper by me, @dbamman.bsky.social, and @ibleaman.bsky.social on cultural NLP: what we want, what we have, and how sociocultural linguistics can clarify things.
Website: naitian.org/culture-not-...
1/n
Check out our new preprint ๐๐๐ง๐ฌ๐จ๐ซ๐๐๐๐.
We use a robust decomposition of the gradient tensors into low-rank + sparse parts to reduce optimizer memory for Neural Operators by up to ๐๐%, while matching the performance of Adam, even on turbulent NavierโStokes (Re 10e5).
PhD student, Srishti Yadav and her collaborators, out with new, interdisciplinary work๐
Check out our new paper led by @srishtiy.bsky.social and @nolauren.bsky.social! This work brings together computer vision, cultural theory, semiotics, and visual studies to provide new tools and perspectives for the study of ~culture~ in VLMs.
A delight to work with great colleagues to bring theory around visual culture and cultural studies to how we think about visual language models.
This work was an amazing collaboration with @nolauren.bsky.social @mariaa.bsky.social @taylor-arnold.bsky.social @jiaangli.bsky.social Siddhesh Pawar, Antonia Karamolegkou, @scfrank.bsky.social @zhaochongan.bsky.social Negar Rostamzadeh, @danielhers.bsky.social @serge.belongie.com Ekaterina Shutova
We find that decades of visual cultural studies offer powerful ways to decode cultural meaning in images!! Rather than proposing yet another benchmark, our goal with this paper was to revisit and re-contextualize foundational theories of culture so that it can pave way for more inclusive frameworks.
We then propose 5 frameworks to evaluate cultures in VLMs:
1๏ธโฃ Processual Grounding - who defines culture?
2๏ธโฃ Material Culture - what is represented?
3๏ธโฃ Symbolic Encoding - how is meaning layered?
4๏ธโฃ Contextual Interpretation - who understands and frames meaning?
5๏ธโฃ Temporality -when is culture situated?
In this paper, we call for integrating methods from 3 fields :
๐ Cultural Studies โ how values, beliefs & identities are shaped through cultural forms like images.
๐ Semiotics โ how signs & symbols convey meaning
๐จ Visual Studies โ how visuals communicate across time & place
Modern Vision-Language Models (VLMs) often fail at cultural understanding. But culture isnโt just recognizing things like food, clothes, rituals etc. It's how meaning is made and understood; it also about symbolism, context, and how these things evolve over time.
Paper title "Cultural Evaluations of Vision-Language Models Have a Lot to Learn from Cultural Theory"
I am excited to announce our latest work ๐ "Cultural Evaluations of Vision-Language Models Have a Lot to Learn from Cultural Theory". We review recent works on culture in VLMs and argue for deeper grounding in cultural theory to enable more inclusive evaluations.
Paper ๐: arxiv.org/pdf/2505.22793
This morning at P1 a handful of lucky of lab members got to see the telescope while centre secretary Bjรถrg had the dome open for a building tour ๐ญ (1/7)
๐New Preprint๐
Can Multimodal Retrieval Enhance Cultural Awareness in Vision-Language Models?
Excited to introduce RAVENEA, a new benchmark aimed at evaluating cultural understanding in VLMs through RAG.
arxiv.org/abs/2505.14462
More details:๐
When you have a lot of work before the deadline push, you keep thinking of others things (distractions) youโd like to do. The day you get free, those things suddenly donโt seem important anymore. And kind of miss work! ๐
This is amazing!! I saw that dataset original webpage was being archived this month. I was wondering whatโll happen to this data.
Screenshot of the dataset viewer on the Hugging Face Hub. Shows a set of metadata for the newspaper navigator dataset. It also has previews of a few rows showing images alongside metadata columns.
๐๏ธ Just released a Parquet version of the Newspaper Navigator dataset on @hf.co!
- 3M+ visual elements from historic US newspapers โ photos, maps, cartoons, OCR + metadata.
- Parquet = fast filters, easier analysis.
- Great for ML + cultural research.
๐ huggingface.co/datasets/big...
We work under this telescope and sometimes get to visit it!