sury
@Surybang
Machine learning education is broken, especially for those who aspire to start solving real-world problems at a company. Most classes, courses, and books start with a dataset and show you how to train a model. dataset → model This is, at best, 5% of the work you'll need to…
Google has made us incapable of memorizing complex things by reducing memorization to search queries. LLMs will make us incapable of solving complex problems by reducing problem-solving skills to writing prompts.
Introducing bitnet.cpp: A blazing-fast open-source 1-bit LLM inference framework that runs directly on CPUs. You can now run 100B parameter models on local x86 CPU devices with up to 6x speed improvements and 82% less energy consumption!
day ???/??? ascii art generation i have successfully added a performant terminal rendering pipeline! frames are rendered as and when they are decoded now (what you see in the video is real time) here's the opening to Serial Experiments Lain on wezterm. 60 fps, locked. with…
remember the GPT Store? a few months ago there was some excitement around "prompt app stores", where gifted prompt-writers could make money by writing magical system prompts and packaging them as "apps" our new research shows these prompts can be easily recovered by just asking…
Check out my new sheets
Huge congrats to @AIatMeta on the Llama 3.1 release! Few notes: Today, with the 405B model release, is the first time that a frontier-capability LLM is available to everyone to work with and build on. The model appears to be GPT-4 / Claude 3.5 Sonnet grade and the weights are…
idk what’s worse, that crowdstrike has kernel access to basically every system in the world or that airlines run on windows server