Skip to content

AI Training

The AI Training page lets you collect, curate, and export training data pairs for fine-tuning language models.

Sidebar item: AI Training Route: /training

What Are Training Pairs?

A training pair consists of a user prompt and a model response. By collecting high-quality pairs, you build a dataset that can be used to fine-tune a model to behave the way you want.

How Pairs Are Collected

Training pairs are automatically captured from your interactions with the Canvas AI feature in the Tool Workshop. When the AI generates tool code, drafts improvements, or provides insights, each interaction becomes a training pair stored in IndexedDB.

Curating Data

The main interface shows a table of all collected training pairs. For each pair you can:

Accept or Reject

  • Accept — Mark the pair as high-quality training data
  • Reject — Mark it as unsuitable (bad output, hallucination, etc.)

Edit Responses

Click on any response to edit it. This is useful when the model's output was close but needs correction — you get the right answer without starting from scratch.

Tagging

Add tags to organize pairs by topic, quality level, or any custom category. Tags help you filter and export specific subsets.

Bulk Operations

Select multiple pairs to:

  • Bulk accept/reject — Set status for many pairs at once
  • Bulk add tags — Apply a tag to all selected pairs
  • Delete — Remove unwanted pairs

Filtering

Filter the training pair list by:

  • Status — Accepted, rejected, or unreviewed
  • Tags — Show only pairs with specific tags
  • Search — Free-text search across prompts and responses

Exporting

Select the pairs you want to export (or use filters to narrow down), then click Export. The exported dataset can be used with fine-tuning tools and frameworks.

Statistics

The page header shows aggregate stats:

  • Total pairs collected
  • Accepted vs. rejected counts
  • Tag distribution

Tips

  • Quality over quantity — A small dataset of carefully curated pairs produces better fine-tuning results than a large noisy dataset.
  • Edit, don't discard — If a response is 80% correct, editing it is more efficient than regenerating from scratch.
  • Use tags strategically — Tag by capability (e.g., "code-gen", "api-tools", "explanation") to create focused training subsets.

Released under the Apache 2.0 License.