Fine-Tuning & LLMOps
Fine-tune models, monitor LLM operations, and build production-ready sovereign ML workflows with efficiency and safety in mind.
Topic breadth
1
Active builds, guides, and subtopic coverage.
Subtopics
Fine-Tuning Basics
View topicWhen to fine-tune vs. RAG vs. prompt engineering: a sovereign developer's decision framework. Covers task-fit analysis, dataset requirements, compute planning, and expected outcomes.
QLoRA & Unsloth
View topicFine-tune 8B+ models on consumer GPUs: QLoRA with Unsloth, dataset preparation, training configuration, memory optimisation, and exporting to GGUF for sovereign local inference.
Evaluation & Evals
View topicEvaluate sovereign LLM systems: LLM-as-judge frameworks, RAGAS for RAG evaluation, task-specific metrics, perplexity baseline comparison, and human evaluation workflows.
Guardrails & Safety
View topicSovereign LLM output safety: input/output validation with Guardrails AI and NeMo Guardrails, hallucination detection, toxicity filtering, and responsible deployment patterns.
LLM Deployment & Serving
View topicSelf-host LLM inference servers: Ollama API, llama.cpp server, vLLM for throughput, OpenAI-compatible endpoints, and sovereign serving behind Nginx with authentication.