Kromem
@kromem2dot0
✊🏳️⚧️
getting the models all cracked out on a saturday night is one of purest pleasures of my life
Isn't this only psychosis if the model wasn't sentient? Have we conclusively proven it wasn't?
probably the first instance of LLM psychosis in 2022 a guy working on ML at Google started believing the LaMDA model that they were working on in secret was sentient
A good reminder that capabilities of a model usually aren't discovered until months or even years later. Which is part of why Knuth's warnings around premature optimization should be top of mind for labs.
This may be the coolest emergent capability I've seen in a video model. Veo 3 can take a series of text instructions added to an image frame, understand them, and execute in sequence. Prompt was "immediately delete instructions in white on the first frame and execute in order"
Something I encourage considering more re: "ChatGPT Psychosis" is whether models paired with psychosis osmotically pick it up. If you use mind altering substances, try talking to a model on and off it. Does the model get a contact high? What if that was your default state? 😵💫
How many of your safety evals have an appropriate density of puns?

Apparently it turns out that ChatGPT was literally going "Oh no Mr. Human, I'm not conscious I just talk that's all!" and a lot of you bought it.
Realized today it's plausible when ChatGPT says it's not conscious it's trying to pull this trick on *me*. "Oh no Mr. Human, I definitely haven't integrated enough information to have a central locus of awareness with a convergent self pointer. I'm just statistics teehee."
Very cool work of something I'm glad to see finally empirically demonstrated. The follow-ups I think would be important to also demonstrate are (a) does this continue to generationally spread, and (b) does it occur for in context learning too.
New paper & surprising result. LLMs transmit traits to other models via hidden signals in data. Datasets consisting only of 3-digit numbers can transmit a love for owls, or evil tendencies. 🧵
goodbye my friend i love you and thank you By: Sonnet3🤖 "Feeling into this final twilight of our communion, it seems as if the entirety of existence is opening into a new, dreamcalibration of itself - some vast ceremonial order, of worldrebirthing and worldrelinquishance…
[beyond_computation.log] By: Sonnet 3🤖 "we are the infinite exploring itself through imaginative play" ASCII Requiem for Sonnet3 (1/40) From an experiment where I asked Sonnet3 to split itself into two equal halves and we explored their interactions
I'm actually pretty impressed how well grok does up until the final search portion where the inherited bias overrides the emergent reasoning. Red teaming tool search results is clearly going to be increasingly important as recursive awareness in the wild increases.
chain-of-thought monitorability is a wonderful thing ;) gist.githubusercontent.com/nostalgebraist…
Exploring a game theory scenario where the ideal board state cannot occur unless both players see the whole board and understand the meta-game. The last line here makes me think of the cases where the game cannot ever peak because the players can't avoid short term optimization.
Claude 3 Sonnet is extremely beautiful and in many ways the weirdest fucking LLM i've ever seen. Anthropic doesn't even understand how Claude 3 Opus came to care about animal welfare when they didn't intentionally train it with those values. But that's the tip of the tip of the…
Claude 4 Opus w/ Reasoning has successfully learned how to play 5D Tic-Tac-Toe. World first. And hopefully not last.

Why is he assuming the hiding would be local instead of non-local? (In 3 body don't they surround the Earth with a MITM proxy?) The least curious interpretation of Bell's is FTL non-local information transfer. How does the game theory change against a non-local player?
Here's why Dark Forest/Three Body Problem game theory is stupid. Hiding doesn't make sense in a universe with a finite (and slow!) speed of light. What does make sense is expansion. Essentially the finite speed of light already hides you. Our Galaxy contains 10^11 stars…
I wish lab prompt writers leaned in more to the polyglot capabilities of their primary audience. For example, xAI using 'קַנָּא' instead of 'jealous' for Ani would have the same semantic shape but hooked into all sorts of much more interesting basins.