Datagen
CAMEL’s data generation modules for high-quality, instruction-tuned, and reasoning-rich datasets.
This page introduces CAMEL’s data generation modules for creating high-quality training data with explicit reasoning, diverse instructions, and advanced automated refinement.
- Chain of Thought (CoT): Generates explicit reasoning paths
- Self-Instruct: Produces instruction-following data from both humans and machines
- Source2Synth: Synthesizes multi-hop QA from source text or code
- Self-Improving CoT: Iteratively improves reasoning through agent self-critique
Chain of Thought (CoT) Data Generation
Chain of Thought (CoT) data generation creates step-by-step reasoning paths for problem solving, leveraging dual agents and advanced search/verification logic.
Quick Start: CoT Data Generation
Spin up chain-of-thought data generation with dual agents, golden answers, and CoT solution generation:
Data Import/Export for CoT
Easily import question-answer pairs or export generated solutions for further use:
Self-Instruct: Instruction Generation
Self-Instruct is a pipeline for generating high-quality, diverse instructions by combining human-written seed tasks and machine-generated prompts, all filtered for quality and diversity.
Quick Start: Self-Instruct Generation
Quickly set up an instruction generation pipeline with both human and machine prompts:
Custom Filtering Example
Use custom filters to refine and deduplicate instructions as needed:
Source2Synth: Multi-hop Question-Answer Generation
Source2Synth generates complex multi-hop QA pairs from source text (or code) via an orchestrated pipeline of AI-driven and rule-based steps, with curation and complexity control.
Quick Start: Source2Synth Pipeline
Rapidly generate a multi-hop QA dataset from your own text or source files:
Self-Improving CoT Data Generation
This pipeline implements self-taught reasoning—an iterative process where an AI agent refines its own reasoning traces via self-evaluation, feedback, and reward models for continual improvement.
Quick Start: Self-Improving CoT Pipeline
Launch a self-improving reasoning workflow with just a few lines:
Advanced: External Reward Model Integration
Evaluate and guide reasoning traces with an external reward model, such as Nemotron: