Hugging Face details how it used its new tool, Skills, to fine tune an LLM using Claude, including for writing scripts, submitting jobs to cloud GPUs, and more
We gave Claude the ability to fine-tune language models using a new tool called Hugging Face Skills. X: @ben_burtenshaw , @donvito , @ben_burtenshaw , @arig23498 , and @ben_burtenshaw . Forums: r/LocalLLaMA X: Ben Burtenshaw / @ben_burtenshaw : We used Claude Code to train open LLMs. Check out the tutorial. basically, we plugged HF skills into claude code and it was able to train LLMs end-to-end. Best thing, this works on all major coding agents: Codex, Cursor, and Gemini CLI. - You tell the agent to fine-tune a model [image] Melvin Vivas / @donvito : claude code fine-tuning a model 🔥 https://huggingface.co/... [image] Ben Burtenshaw / @ben_burtenshaw : just to clarify, yes! claude code is contributing to open source ai. Aritra / @arig23498 : The idea is WILD. One idea would be to ask the Claude Code to figure out the hyperparameters that best train a regime of models. Another one would be to work on distillation of bigger models. The ideas are endless here (really cool work) Ben Burtenshaw / @ben_burtenshaw : @ivibecode We actually have a benchmarking skill, that evaluates models. We're going to do another post on that when it's finished. Forums: r/LocalLLaMA : We Got Claude to Fine-Tune an Open Source LLM