Skip to content

Overview

This directory contains practical examples demonstrating various use cases of the distil labs platform. Each example is contained in its own directory with complete code and documentation. To see more examples visit our examples GitHub repository

This tutorial walks you through the full distil labs workflow from installation to deployment, using a PII redaction task as an example.

This example demonstrates how to fine-tune a small language model (SLM) for a RAG system.

Text classification: training and deployment

Section titled “Text classification: training and deployment”

This example demonstrates how to fine-tune a small language model (SLM) for text classification.

Question Answering: training and deployment for open book QA

Section titled “Question Answering: training and deployment for open book QA”

This example demonstrates how to fine-tune a small language model (SLM) for open-book question answering. In the Open-book QA setting, the LLM can refer to external sources of information (e.g., contexts gathered by the retriever) to answer given questions.