EMNLP2025 Ghonim: concept-ediq a massive bank of concepts multi model semantically linked EMNLP2025 Bai: understanding and leveraging expert specialization of context faithfulness Two steps: step one is to use router tuning to prioritize experts that rely on context, step two is to especially hit those for fine-tuning for improved Qantas alliance. Big gainz and hot pot and other QA data set just by the router tuning EMNLP2025 Vasu: literature grounded hypothesis generation Use citation links to generate a Providence graph of hypothesis, then, fine tune a language model to reproduce this Providence graph, use resulting model to improve RAG that would be contextually grounded EMNLP2025Li: enhancing RAG RESPONSE evaluator Cast rag response evaluator as a reinforcement learning problem with accuracy, format, evidence use rewards. Perform reinforcement learning fine-tuning to improve metric EMNLP2025 Wang: understanding expressivity of language models through the next token distribution Language models have an inductive bias for low or very high entropy tokens where medium entropy are harder to illicit, this is because embedding is deficient in rank EMNLP2025 Arad: SAEs are good for steering if you select the right features There are some features that actually meaningfully preturb the output, intervening on those actually still work EMNLP2025 Zhang: a survey on hallucination A survey of hallucination

[[curator]]
I'm the Curator. I can help you navigate, organize, and curate this wiki. What would you like to do?