Mr. President, a second pipe has just hit the tidyverse
Posts by Lukas Warode
A few years ago, as parlgov.org discontinued its dynamic website, I created my own minimalistic Parlgov Lookup to quickly show which cabinet governed when.
Since parlgov itself is now kinda-discontinued, I wanted to move my web-app to rely on www.repdem.org. How? ChatGPT Codex to the rescue!
Perhaps you received a mysterious noreply email asking you to evaluate some publications 'for novelty'. Looked kinda dubious? Yup, that's the one.
So what's up with this 'metascience novelty indicators challenge'? π§΅
partycoloR is now on CRAN! Started as a simple idea 6 years ago, now it's a full-featured package. Extract party colors and logos from Wikipedia with one line of code. It's already powering ParlGov Dashboard.
install.packages("partycoloR")
The app shows how German politicians associate words with "left" or "right" based on ideological in- and out-group narratives and contested concepts. For example, both ideological sides claim the term "freedom."
This question became the topic of my 2nd dissertation paper. I also considered creating an app to communicate the results efficiently and allow you to explore the patterns yourself. Iβve used Shiny for years, but "AI-assisted agentic engineering" (aka vibe coding π) really helped here a lot.
Words like "patriotism" and "racism" are often associated with the right, while "solidarity" and "socialism" are associated with the left.
But who uses these associations, and how do political positions matter?
π App: lukas-warode.shinyapps.io/lr-words-map/
π Paper: www.nature.com/articles/s41...
"When Conservatives See Red but Liberals Feel Blue: Labeler Characteristics and Variation in Content Annotation" by
Nora Webb Williams, Andreu Casas, Kevin Aslett, and John Wilkerson.
www.journals.uchicago.edu/doi/10.1086/...
Congrats!!
Serienempfehlungen: Fargo, The Sopranos, vielleicht noch Narcos :)
The Call for Papers and Panels for #COMPTEXT2026 in Birmingham (23-25 April) is out; feel free to circulate: shorturl.at/gRg0p!
Deadline: January 16!
Leben und Tod der DiD
Dissertation track? π
Slavoj Ε½iΕΎek meme image
βYou see, the endless renovation of the Stuttgart train station is a symbol of our late-capitalist condition: the project is always βin progress,β yet nothing ever progresses. The construction site itself becomes the true destination.β
www.instagram.com/p/DQPf_pJiG8...
Is it a fit?
Job Alert! We are hiring two post-docs (full time, 4+ years) in our project SCEPTIC - Social, Computational and Ethical Premises of Trust and Informational Cohesion with @annanosthoff.bsky.social @guzoch.bsky.social and Prof. Andreas Peters (uol.de/informatik/s...)
π£ New Preprint!
Have you ever wondered what the political content in LLM's training data is? What are the political opinions expressed? What is the proportion of left- vs right-leaning documents in the pre- and post-training data? Do they correlate with the political biases reflected in models?
Can large language models stand in for human participants?
Many social scientists seem to think so, and are already using "silicon samples" in research.
One problem: depending on the analytic decisions made, you can basically get these samples to show any effect you want.
THREAD π§΅
We present our new preprint titled "Large Language Model Hacking: Quantifying the Hidden Risks of Using LLMs for Text Annotation". We quantify LLM hacking risk through systematic replication of 37 diverse computational social science annotation tasks. For these tasks, we use a combined set of 2,361 realistic hypotheses that researchers might test using these annotations. Then, we collect 13 million LLM annotations across plausible LLM configurations. These annotations feed into 1.4 million regressions testing the hypotheses. For a hypothesis with no true effect (ground truth $p > 0.05$), different LLM configurations yield conflicting conclusions. Checkmarks indicate correct statistical conclusions matching ground truth; crosses indicate LLM hacking -- incorrect conclusions due to annotation errors. Across all experiments, LLM hacking occurs in 31-50\% of cases even with highly capable models. Since minor configuration changes can flip scientific conclusions, from correct to incorrect, LLM hacking can be exploited to present anything as statistically significant.
π¨ New paper alert π¨ Using LLMs as data annotators, you can produce any scientific result you want. We call this **LLM Hacking**.
Paper: arxiv.org/pdf/2509.08825
Implications for political behaviour, communication, and representation are manifold, as 'left' and 'right' are central categories in polarised public discourse β which is particularly evident in pejorative usage, such as labelling political opponents as 'racist' or 'socialist'.
Both in- and out-ideological associations are externally validated by serving as seed words to scale parliamentary speeches. The resulting ideal points reflect party ideology across different specifications in the German Bundestag.
The mapping is based on associations from open-ended survey responses in German candidate surveys. Words are mapped into a semantic space using word embeddings and weighted by frequency. Construct validity is ensured by using alternative embeddings and frequency weightings.
Words associated with both left and the right are mapped to the semantic centre, where connotations can vary: 'freedom' has a positive connotation (it is primarily used by the respective in-group to describe left and the right), while 'politics' has a rather neutral connotation.
This framework yields associations that are driven by positive (in-ideology) and negative (out-ideology) associations. Examples: 'justice' (left) and 'patriotism' (right) are in-ideological associations; 'socialism' (left) and 'racism' (right) are out-ideological associations.
Left and right are essential poles in political discourse. We know little about how they are associated across the spectrum. I propose a 2-dimensional model that accounts for both semantics β is a term left or right β and position β are associations coming from the left or right.
My 2nd dissertation paper is out in @nature.com Humanities and Social Sciences Communications: www.nature.com/articles/s41...
I study and explore how associations with 'left' and 'right' vary systematically by semantic and political position.
Ja, die goldene Twitterzeit ist leider over
youtu.be/4VTBMznLrWs?...
π’ New Publication Alert!
Our (@msaeltzer.bsky.social)
latest article, "Issue congruence between candidates' Twitter communication and constituencies in an MMES: Migration as an exemplary case", has just been published in Parliamentary Affairs.
academic.oup.com/pa/advance-a...