Alexander Holden Miller
@alex_h_miller
Research Engineering Manager at FAIR @MetaAI
1/7 For the past decade, our team at Meta Reality Labs (previously CTRL-labs) has been dedicated to developing a neuromotor interface. Our goal is to address the Human Computer Interaction challenge of providing effortless, intuitive, and efficient input to computers.
Hiring! We're looking to fill contractor Research Engineer roles in New York City to work with us in FAIR on AI Research Agents. If that sounds fun, please fill out the expression of interest here: forms.gle/7m4fVqLXY5GwuL…
We’re honored to share that Learning-Rate-Free Learning by D-Adaptation, a paper by Meta AI research scientist @aaron_defazio & @konstmish was selected as an #ICML2023 Outstanding Paper! More details and a link to the paper ⬇️
Learning Rate Free Adam? @konstmish and I have developed D-Adaptation, a method for automatically adjusting the learning rate during training that ACTUALLY WORKS IN PRACTICE on a broad range of deep learning problems. arxiv.org/abs/2301.07733 1/2
"over 200k open models, 5k new ones just last week, 50k open datasets, and 100k applications" amazing work happening at 🤗 such an elegant testimony in support of open source and open science from @ClementDelangue
This is my 5-minute testimony before the US Congress! Open science and open source AI distribute economic gains by enabling hundreds of thousands of small companies and startups to build with AI. It fosters innovation, and fair competition between all. Thanks to ethical…
We've just released MusicGen, and there is a @huggingface demo now, here is a thread about me playing with it just right now. huggingface.co/spaces/faceboo… A 🧵👇
🚨 New work: BlenderBot 3x 🚨 - Public data release & analysis of 6M chat interactions. - Learns by conversing with people in the real world: training on this data improves BB3 from 85.3% → 94.4% good messages. paper: arxiv.org/abs/2306.04707 project: parl.ai/projects/bb3x/
New paper on scaling language models to sequences of a million bytes! MegaByte splits long byte sequences into fixed-size patches (analogous to tokens), then runs a large model between the patches, and a small model to predict each patch byte-by-byte. 1/
MEGABYTE: Predicting Million-byte Sequences with Multiscale Transformers abs: arxiv.org/abs/2305.07185 paper page: huggingface.co/papers/2305.07…
@Meta AI's CICERO algorithm — which negotiates and build trust with humans to perform in the top decile at the game of Diplomacy — is (in my view) the most astounding #AI feat yet. Hear all about it from @alex_h_miller. #SuperDataScience #NLP Watch here: superdatascience.com/663
Huge pleasure to be a guest of Jon's! He's a great interviewer that strikes the perfect balance between indulging curiosity and staying focused on content and doesn't waste your time on chit-chat. Check out the rest of his "Super Data Science" podcast!
@Meta AI's CICERO algorithm — which negotiates and build trust with humans to perform in the top decile at the game of Diplomacy — is (in my view) the most astounding A.I. feat yet. And now I've interviewed @alex_h_miller about it :) Alex's episode will likely be #663 and air on…
Today we release LLaMA, 4 foundation models ranging from 7B to 65B parameters. LLaMA-13B outperforms OPT and GPT-3 175B on most benchmarks. LLaMA-65B is competitive with Chinchilla 70B and PaLM 540B. The weights for all models are open and available at research.facebook.com/publications/l… 1/n
Today we're publicly releasing LLaMA, a state-of-the-art foundational LLM, as part of our ongoing commitment to open science, transparency and democratized access to new research. Learn more & request access ➡️ bit.ly/3SoXdQE
Using our HyperTree Proof Search method we created a neural theorem solver that was able to solve 10 International Math Olympiad problems — 5x any previous AI system. More in this post ➡️ bit.ly/3vBnu3E
Meta AI’s @alex_h_miller speaks to CICERO's ability to perform far beyond today's 'scripted' AI agents. #CICERObyMetaAI uses purposeful, intentional language to understand and interact w/ Diplomacy players to achieve shared goals.
After having had some time to digest, I actually think of all the big AI things to drop in the last month, Cicero is the most impressive. It's the first time I've seen LLMs used in a way where there really seems to be a clear meaning or intention behind the utterances.
We're the Meta AI research team behind CICERO, the first AI agent to achieve human-level performance in the game Diplomacy. We’ll be answering your questions on December 8th starting at 10am PT. Ask us anything! #ai #diplomacy reddit.com/r/MachineLearn…
See what the humans behind #CICERObyMetaAI have to say about the breakthrough AI agent they built: @polynoamial, @em_dinan, @alex_h_miller, @adamlerer, @anton_bakhtin, @ml_perception and @joespeez. For more on CICERO, read our blog: bit.ly/3iIhxyB