Capture Runs. Curate Data. Fine-Tune Models. Ship Better Agents.
Learn the full agent improvement lifecycle: capture good and bad agent runs, curate a training dataset, fine-tune or distill a model on Nebius, and redeploy behind the same workflow. All running on Nebius Serverless + Token Factory — no infrastructure changes needed to go from prototype to fine-tuned production model.
Jump to Step-by-Step GuideAI engineers, ML-adjacent developers, teams with agents already in production
Continuous agent improvement loop from data capture to fine-tuned deployment
"We fine-tuned a 7B model that outperforms GPT-4 on our specific workflow — and it runs 10x cheaper"
A data capture pipeline that logs agent runs with quality labels
A curated fine-tuning dataset from real agent interactions
A fine-tuned model deployed on Token Factory behind your existing workflow
Why fine-tuning matters and how the data → train → deploy loop works
Set up run logging, label outputs, and build a training dataset
Fine-tune a model on Nebius and deploy it behind your existing workflow
Advanced patterns: model distillation, A/B testing, and continuous improvement
Follow these steps during the workshop. Each step includes commands you can copy, tips from our mentors, and a checkpoint to verify before moving on.
Deploy OpenClaw on Nebius Serverless with Token Factory — or use your existing setup from Workshop 1.
# Verify existing endpoint (if you have one)nebius msp serverless v1alpha1 endpoint get-by-name --name openclaw-agent# Or deploy fresh (see Workshop 1 for full setup)nebius msp serverless v1alpha1 endpoint create \--name openclaw-lifecycle \--container-image openclaw:latest \--container-template-resources-platform cpu-d3 \--container-template-resources-preset 4vcpu-16gb \--port 8080 \--username admin \--password "$(openssl rand -hex 32)" \--network-id <your-network-id> \--parent-id <your-project-id>
Your OpenClaw agent is running and processing requests on Nebius Serverless.
Instrument your agent to log every run — inputs, outputs, tool calls, and timing — in a structured format.
Running your agent produces a JSONL log file with full interaction details for each run.
Go through captured runs and label them as good, bad, or needs-improvement. This is your training signal.
You have 30+ labeled runs with quality scores and notes on what went right/wrong.
Transform your labeled runs into a fine-tuning dataset in chat completion format.
You have a training JSONL file and an eval JSONL file, both in chat completion format.
Launch a fine-tuning job on Nebius using your curated dataset. We'll fine-tune a Llama model to specialize on your workflow.
Your fine-tuning job completes and the eval shows improvement over the base model.
Deploy your fine-tuned model on Token Factory and swap it into your existing agent workflow — no code changes needed.
Your agent is running on the fine-tuned model and producing better results than the base model.
Set up monitoring to track whether your fine-tuned model actually improves outcomes over time, and plan the next iteration.
# Check endpoint metricsnebius msp serverless v1alpha1 endpoint get $ENDPOINT_ID# View recent logsnebius msp serverless v1alpha1 endpoint logs $ENDPOINT_ID
You have monitoring in place, a documented process, and a plan for the next fine-tuning cycle.
RSVP required. Spots are limited since we provide hands-on support for every attendee.
Register Now