Axolotl
@axolotl_ai
Axolotl is the premier open source LLM fine tuning framework. find us on discord https://discord.gg/cq8QGrh9mC
Hacker house was in full force last week for @aiDotEngineer with events with @runpod_io plus collab with @googlecloud... sailing with team + @capetorch visits @SHACK15sf ... ⛵️🔥 😊
Wrapping up great week building and bonding with @axolotl_ai
🚨 Introducing the Axolotl-LLM Compressor integration, designed to make fine-tuning sparse models easier and more efficient than ever! Now you can fine-tune sparse models for specific data while preserving their sparse structure and recovering any accuracy lost during pruning.…
Training qwe3 with qwen2.5 template works just fine. Way easier to make everything work, and just one line change in @axolotl_ai
Looking to fine-tune models? Meet @axolotl_ai 💚 an open-source tool that simplifies the entire fine-tuning pipeline. Learn how to get started with our tutorial covering how to fine-tune Llama 3 🦙 koyeb.com/tutorials/usin…
Magistral Small is an open-weight model, and is available for self-deployment under the Apache 2.0 license. You can use it with the following frameworks: - vllm (recommended): huggingface.co/mistralai/Magi… The community has prepared quantized versions of the model that can be used…
Using @googlecloud 🤝 @axolotl_ai can help you streamline your large Multimodal finetuning workflows.
Deploy Axolotl with @googlecloud for your production workloads with simple configuration based orchestration.
Using @googlecloud 🤝 @axolotl_ai can help you streamline your large Multimodal finetuning workflows.
Spot the runpod truck in sf and get $250 in gpu credits Snap a pic + tag @runpod_io and we'll send you a credit code 🤝 Hint: it's right by @aiDotEngineer
Model Merging in Pre-training of Large Language Models "We present the Pre-trained Model Averaging (PMA) strategy, a novel framework for model merging during LLM pre-training. Through extensive experiments across model scales (from millions to over 100B parameters), we…