Nikola Jurkovic
@nikolaj2030
AI safety researcher
I'm genuinely curious what would the public reaction be if a lab suddenly posted: "We did it. AGI happened. But it escaped. All we know is it dumped itself off the server and could be anywhere." Like… is there even a Protocol One for that? Someone explain what happens next.
the openai IMO news hit me pretty heavy this weekend i'm still in the acute phase of the impact, i think i consider myself a professional mathematician (a characterization some actual professional mathematicians might take issue with, but my party my rules) and i don't think i…
I haven't written about this much or thought it through in detail, but here are a few aspects that go into my backdrop model: (1) especially in the long-term technological limit, I expect human labor to be wildly uncompetitive for basically any task relative to what advanced…
A simple AGI safety technique: AI’s thoughts are in plain English, just read them We know it works, with OK (not perfect) transparency! The risk is fragility: RL training, new architectures, etc threaten transparency Experts from many orgs agree we should try to preserve it:…