MilaNLP
@MilaNLProc
The Milan Natural Language Processing Group #NLProc #ML #AI
#TBT #NLProc 'Ecological Fallacy in Annotation' by Orlikowski et al. (2023) posits that sociodemographics in annotator models have minimal impact on performance. aclanthology.org/2023.acl-short…
📖For this Thursday's Reading Group, Serena Pugliese presented "Emergent social conventions and collective bias in LLM populations" by Ariel Flint Ashery et al. (2025) Paper: science.org/doi/10.1126/sc… #NLProc

#MemoryMonday #NLProc 'Predicting News Headline Popularity' by Lamprinidis, Hardt, Dirk Hovy (2018) shows neural networks perform similar to Logistic Regression in prediction. aclanthology.org/D18-1068
#TBT #NLProc : Bergman et al.'s 'Guiding the Release of Safer E2E Conversational AI through Value Sensitive Design' explores AI launch with a value-sensitive lens. aclanthology.org/2022.sigdial-1…
🎉 The @MilaNLProc lab is excited to present 15 papers and 1 tutorial at #ACL2025 & workshops! Grateful to all our amazing collaborators, see everyone in Vienna! 🚀



🥁 #WOAH2025 Speakers🥁 We are excited to announce an amazing panel of 3 speakers who will discuss our theme "Harms Beyond Hate Speech" 🗣️ @CordeliaMoore 🗣️ @fvancesco 🗣️ @katejsim Full details: workshopononlineabuse.com/keynotes.html 🤩 See you on August 1!
#MemoryMonday #NLProc 'The Empty Signifier Problem: Towards Clearer Paradigms for Operationalising 'Alignment' in Large Language Models' by Kirk et al. (2023) presents a novel framework for unpacking 'alignment' in LLMs, promoting definitional clarity. arxiv.org/abs/2310.02457
🎓For today's lab seminar, we had the pleasure to host @MiriamSchirmer with her presentation on Measuring and Reducing the Psychological Impact of Online Harm and @tpimentelms with How Much Does Tokenisation Impact Language Models? #NLProc #onlineharms #tokenisation


#TBT #NLProc 'Tonneau et al.'s (2024) 'From Languages to Geographies'' reveals geo-cultural bias in hate speech data, underlining the need for improved representation. #Research #BiasInData aclanthology.org/2024.woah-1.23
📖For this Thursday's Reading Group, Pranav presented "Research Borderlands: Analysing Writing Across Research Cultures" by Bhatt et al. (2025) Paper: arxiv.org/pdf/2506.00784 #NLProc

#MemoryMonday #NLProc 'Exploring challenges in Zero-shot Cross-lingual Hate Speech Detection, @debora_nozza (2021) reveals how current models may inaccurately label non-hateful, language-specific interjections as hate speech signals.' aclanthology.org/2021.acl-short…
Last week we held our 1st MilaNLP retreat by beautiful Lago Maggiore! ⛰️🌊 We shared research ideas, stories (academic & beyond), and amazing food. It was a great time to connect outside of the usual lab working days, and most importantly, strengthen our bonds as a team. #NLProc



#TBT #NLProc 'Emotion Analysis in NLP: Trends, Gaps & Future' by Plaza-del-Arco et al. (2024) analyses key NLP papers, spotting major gaps in emotion analysis including terminology standardization and demographic aspects. aclanthology.org/2024.lrec-main…
#MemoryMonday #NLProc 'Language Invariant Properties in NLP' by Bianchi, @debora_nozza, Dirk Hovy (2022) explores text transformation constants, studying translation & paraphrasing effects on sentiment & traits. #linguistics aclanthology.org/2022.nlppower-…
🎓For today's lab seminar, we've had the pleasure to listen to @pietro_lesci 's presentation on "The hidden cost of split: Tokenization bias in language models". #NLProc

#TBT #NLProc Metrics for What, Metrics for Whom: Assessing Actionability of Bias Evaluation Metrics in NLP" by Delobelle, Attanasio, @deboranozza.bsky.social, Blodgett, and Talat, 2024 highlights the need for clearer, actionable bias measures in NLP. aclanthology.org/2024.emnlp-mai…
📖For this Thursday's Reading Group, Yujie Ren presented "How Does Vision-Language Adaptation Impact the Safety of Vision Language Models?" by Lee et al 2024 Paper: arxiv.org/abs/2410.07571 #NLProc #AISafety

#MemoryMonday #NLProc 'Data-Efficient Strategies for Expanding Hate Speech Detection into Under-Resourced Languages' funnels limited target-language data to fine-tune models & enhance effectiveness. By @paul_rottger et al. aclanthology.org/2022.emnlp-mai…
🎓For today's lab seminar it was a pleasure to have Federico Ruggeri presenting "From data annotation to modelling, the importance of integrating and extracting textual knowledge in hate speech" #NLProc

#TBT #NLProc New study, 'Entropy-based Attention Regularization Frees Unintended Bias Mitigation from Lists' by Attanasio et al. redefines bias reduction in #AI, sans prior term knowledge. #2022Publication aclanthology.org/2022.findings-…