Ricardo Buitrago
@rbuit_
ML at Cartesia AI | CMU
Despite theoretically handling long contexts, existing recurrent models still fall short: they may fail to generalize past the training length. We show a simple and general fix which enables length generalization in up to 256k sequences, with no need to change the architectures!

Tokenization has been the final barrier to truly end-to-end language models. We developed the H-Net: a hierarchical network that replaces tokenization with a dynamic chunking process directly inside the model, automatically discovering and operating over meaningful units of data
I converted one of my favorite talks I've given over the past year into a blog post. "On the Tradeoffs of SSMs and Transformers" (or: tokens are bullshit) In a few days, we'll release what I believe is the next major advance for architectures.
Today, we're excited to share Cartesia Narrations, in public beta. Narrations is a creator tool for narrating long-form content using Sonic 2.0. You can use it to create audiobooks and podcasts, narrate your Substack posts, and more. Some highlights🧵