Canfer Akbulut
@canfer_akbulut
sociotechnical AI research @googledeepmind
super exciting job opportunity! 🥳
Are you interested in exploring questions at the ethical frontier of AI research? If so, then take a look at this new opening in the humanity, ethics and alignment research team: boards.greenhouse.io/deepmind/jobs/… HEART conducts interdisciplinary research to advance safe & beneficial AI.
I'm presenting our work on the Gaps in the Safety Evaluation of Generative AI today at @AIESConf ! We survey the state of safety evaluations and find 3 gaps: the modality gap 📊, the coverage gap 📸, and the context gap 🌐. Find out more in the paper: ojs.aaai.org/index.php/AIES…

I am really beyond words to be able to share with the world what @bakkermichiel, @summerfieldlab, and I and a truly world-class team @GoogleDeepMind worked on for the past 2 years. Out today in @ScienceMagazine, “AI can help humans find common ground in democratic deliberation”
What does it mean for AI to be "too human"? How could such a situation arise? And why does this matter? Check out this excellent paper by @canfer_akbulut (and others) out at @AIESConf today! 🙌 ojs.aaai.org/index.php/AIES…
Have you been thinking about the implications of anthropomorphic AI quite a bit this week? 🤔 We explore the risks of anthropomorphic AI systems in our Ethics of Advanced AI Assistants report. Key insights in thread 💡deepmind.google/discover/blog/…
How do we anticipate and prepare for the impacts of anthropomorphic AI on users and society? We explore this question in our paper, now out in @AIESConf #aies2024: ojs.aaai.org/index.php/AIES…
Have you been thinking about the implications of anthropomorphic AI quite a bit this week? 🤔 We explore the risks of anthropomorphic AI systems in our Ethics of Advanced AI Assistants report. Key insights in thread 💡deepmind.google/discover/blog/…
In a world where users rely on advanced AI assistants for a range of tasks across various domains, when would user trust in the technology be justified? Our @FAccTConference paper explores this question. Join our presentation at 11.35 am this morning! Here are 3⃣ key insights.
Most real-world AI applications involve human-model interaction, yet most current safety evaluations do not. In a new paper with @saffronhuang @_lamaahmad @Manderljung, we argue that we need evaluations which assess human-model interactions for more accurate safety assessments 🧵
More great work from a research team led by our model methodologist and evaluator in chief @weidingerlaura 👏 Here's what we learned during the latest round of @GoogleDeepMind model testing🤖📊
📜 New paper unpacking Google DeepMind’s approach to safety evals for advanced AI models, with lessons learned to support the advancement of similar efforts by other actors in this space. Covers foresight, evaluation design, and the wider ecosystem. arxiv.org/abs/2404.14068
paper's out! 🥳 honored to have contributed to this thorough analysis of the mechanisms that enable persuasive generative AI – hats off to seliem and sasha for their spectacular leadership!
Our new paper on AI persuasion, exploring definitions, harms and mechanisms. Happy to have contributed towards the section on mitigations to avoid harmful persuasion. Some highlights in 🧵 storage.googleapis.com/deepmind-media…
A truly monumental effort led by @IasonGabriel , @Arianna_Manzini , and Geoff Keeling. I led the section on anthropomorphism – many thanks to my lovely co-authors Iason, Arianna, @verena_rieser, and @weidingerlaura and to @merrierm and @mhtessler for their thoughtful edits!
1. What are the ethical and societal implications of advanced AI assistants? What might change in a world with more agentic AI? Our new paper explores these questions: storage.googleapis.com/deepmind-media… It’s the result of a one year research collaboration involving 50+ researchers… a🧵
We’re excited to announce 𝗚𝗲𝗺𝗶𝗻𝗶: @Google’s largest and most capable AI model. Built to be natively multimodal, it can understand and operate across text, code, audio, image and video - and achieves state-of-the-art performance across many tasks. 🧵 dpmd.ai/announcing-gem…
From assisting in healthcare to creating art, generative AI is changing how we live and work. We developed a framework to evaluate its risks at the point of technological capability, human interaction & systemic impact. Here's a snapshot of the work. 🧵 dpmd.ai/EvalsGenAI_TW