Skills

All Skills

Fine-Tuning

Skills tagged with #Fine-Tuning

@ml-explore

swift-mlx-lm

MLX Swift LM - Run LLMs and VLMs on Apple Silicon using MLX. Covers local inference, streaming, wired memory coordination, tool calling, LoRA fine-tuning, embeddings, and model porting.

ml-explore/mlx-swift-lm
18d ago
2940
@cerebrixos-org
MCP

Tuning Engines

Domain-specific LLM fine-tuning — sovereign models trained on your data, zero infrastructure.

mcpgithubaillm
cerebrixos-org/tuning-engines-cli
19d ago
0
@huggingface

trl-training

Train and fine-tune transformer language models using TRL (Transformers Reinforcement Learning). Supports SFT, DPO, GRPO, KTO, RLOO and Reward Model training via CLI commands.

huggingface/trl
18d ago
17.7K0
@LazyAGI

lazyllm-skill

LazyLLM framework for building multi-agent AI applications. Use when task mentioned LazyLLM or AI program for: (1) Flow orchestration - linear, branching, parallel, loop workflows for complex data pipelines, (2) Model fine-tuning and acceleration - finetuning LLMs with LLaMA-Factory/Alpaca-LoRA/Collie and acceleration with vLLM/LMDeploy/LightLLM. Includes comprehensive code examples for all components, (3) RAG systems - knowledge-based QA with document retrieval, vectorization, and generation, (4) Agent development - single/multi-agent systems with tools, memory, planning, and web interfaces.

LazyAGI/LazyLLM
18d ago
3.7K0
@applied-artificial-intelligence

huggingface-transformers

Hugging Face Transformers best practices including model loading, tokenization, fine-tuning workflows, and inference optimization. Use when working with transformer models, fine-tuning LLMs, implementing NLP tasks, or optimizing transformer inference.

applied-artificial-intelligence/claude-code-toolkit+5 more
18d ago
440
@Red-Hat-AI-Innovation-Team

training-hub-guide

Guides users through LLM post-training with Training Hub, including installation, algorithm selection (SFT, OSFT, LoRA), hyperparameter tuning, troubleshooting OOM errors, interpreting loss curves, and leveraging backend-specific features. Use when the user is working with training_hub, fine-tuning language models, asking about SFT/OSFT/LoRA training, or debugging GPU/CUDA training issues.

Red-Hat-AI-Innovation-Team/training_hub
18d ago
650
@brevdev

brev-cli

Manage GPU and CPU cloud instances with the Brev CLI for ML workloads and general compute. Use when users want to create instances, search for GPUs or CPUs, SSH into instances, open editors, copy files, port forward, manage organizations, or work with cloud compute. Supports fine-tuning, reinforcement learning, training, inference, batch processing, and other ML/AI workloads. Trigger keywords - brev, gpu, cpu, instance, create instance, ssh, vram, vcpu, A100, H100, cloud gpu, cloud cpu, remote machine, finetune, fine-tune, RL, RLHF, training, inference, deploy model, serve model, batch job.

brevdev/brev-cli
18d ago
2440
@itsmostafa

mlx

Running and fine-tuning LLMs on Apple Silicon with MLX. Use when working with models locally on Mac, converting Hugging Face models to MLX format, fine-tuning with LoRA/QLoRA on Apple Silicon, or serving models via HTTP API.

itsmostafa/llm-engineering-skills+1 more
3d ago
240
@travisjneuman

ai-ml-development

AI and machine learning development with PyTorch, TensorFlow, and LLM integration. Use when building ML models, training pipelines, fine-tuning LLMs, or implementing AI features.

travisjneuman/.claude+50 more
18d ago
300
@muratcankoylan

book-sft-pipeline

This skill should be used when the user asks to "fine-tune on books", "create SFT dataset", "train style model", "extract ePub text", or mentions style transfer, LoRA training, book segmentation, or author voice replication.

muratcankoylan/Agent-Skills-for-Context-Engineering+5 more
19d ago
13.9K0