This work was conducted at @copenlu.bsky.social under the guidance of my amazing supervisors, @iaugenstein.bsky.social and @apepa.bsky.social.
Posts by Sekh (Sk) Mainul Islam
Overall, this work advances understanding of how LLMs integrate internal and external knowledge by introducing the first systematic framework for multi-step analysis of knowledge interactions via rank-2 subspace disentanglement.
π‘How is the CoT mechanism aligned with the knowledge interaction subspace?
π CoT maintains similar CK alignment compared to standard prompting for all the datasets, and also reduces PK alignment.
π‘ Can we find reasons for hallucinations based on PK-CK interactions?
π The gap between PK and CK is much higher for the examples with hallucinated spans than for the examples with no hallucinated spans across the sequence steps.
π‘ How do individual PK and CK contributions change over the NLE generation steps for different knowledge interactions?
π During most of the NLE generations, the model slightly prioritizes PK.
π‘ How do individual PK and CK contributions change over the NLE generation steps for different knowledge interactions?
π While generating an answer, the model aligns with the CK direction for conflicting examples, while for supportive examples, the model aligns with PK.
πͺ We propose a novel rank-2 projection subspace that disentangles PK and CK contributions more accurately and use it for the first multi-step analysis of knowledge interactions across longer NLE sequences.
π‘ Is a rank-1 projection subspace enough for disentangling PK and CK contributions in all types of knowledge interaction scenarios?
π Different knowledge interactions are poorly captured by the rank-1 projection subspace in LLM model parameter
Prior work has largely examined only single-step generation β typically the final answer, and has modelled PKβCK interaction only as a binary choice in a rank-1 subspace. This overlooks richer forms of interaction, such as complementary or supportive knowledge.
π€ NLEs illustrate the underlying decision-making process of LLMs in a human-readable format and reveal the utilization of PK and CK. Understanding their interaction is key to assessing the grounding of NLEs, yet it remains underexplored.
I am excited to share our new preprint answering this question:
"Multi-Step Knowledge Interaction Analysis via Rank-2 Subspace Disentanglement"
π Paper: arxiv.org/pdf/2511.01706
π» Code: github.com/copenlu/pk-c...
What is the interaction dynamics between Parametric Knowledge (PK) and Context Knowledge (CK) in generating longer Natural Language Explanation (NLE) sequences?
π©βπ¬ Huge thanks to my brilliant co-authors from @copenlu.bsky.social (led by @iaugenstein.bsky.social ) β @nadavb.bsky.social , Siddhesh Pawar, @haeunyu.bsky.social , and @rnv.bsky.social .
@aicentre.dk
π Key Takeaways:
3οΈβ£ Real & Fictional Bias Mitigation: Reduces both real-world stereotypes (e.g., βItalians are reckless driversβ) and fictional associations (e.g., βcitizens of a fictional country have blue skinβ), making it useful for both safety and interpretability research.
π Key Takeaways:
2οΈβ£ Strong Generalization: Works on unseen biases during token-based fine-tuning.
π Key Takeaways:
1οΈβ£ Consistent Bias Elicitation: BiasGym reliably surfaces biases for mechanistic analysis, enabling targeted debiasing without hurting downstream performance.
BiasGym consists of two components:
BiasInject: injects specific biases into the model via token-based fine-tuning while keeping the model frozen.
BiasScope: leverages these injected signals to identify and steer the components responsible for biased behaviour.
π‘ Our Approach: We propose BiasGym, a simple, cost-effective, and generalizable framework for surfacing and mitigating biases in LLMs through controlled bias injection and targeted intervention.
π Problem: Biased behaviour of LLMs is often subtle and non-trivial to isolate, even when deliberately elicited, making systematic analysis and debiasing particularly challenging.
π Excited to share our new preprint: BiasGym: Fantastic LLM Biases and How to Find (and Remove) Them
π Read the paper: arxiv.org/abs/2508.08855