docs/en/concepts/training.mdx
The training feature in CrewAI allows you to train your AI agents using the command-line interface (CLI).
By running the command crewai train -n <n_iterations>, you can specify the number of iterations for the training process.
During training, CrewAI utilizes techniques to optimize the performance of your agents along with human feedback. This helps the agents improve their understanding, decision-making, and problem-solving abilities.
To use the training feature, follow these steps:
crewai train -n <n_iterations> -f <filename.pkl>
To train your crew programmatically, use the following steps:
n_iterations = 2
inputs = {"topic": "CrewAI Training"}
filename = "your_model.pkl"
try:
YourCrewName_Crew().crew().train(
n_iterations=n_iterations,
inputs=inputs,
filename=filename
)
except Exception as e:
raise Exception(f"An error occurred while training the crew: {e}")
CrewAI uses the training artifacts in two ways: during training to incorporate your human feedback, and after training to guide agents with consolidated suggestions.
flowchart TD
A["Start training
CLI: crewai train -n -f
or Python: crew.train(...)"] --> B["Setup training mode
- task.human_input = true
- disable delegation
- init training_data.pkl + trained file"]
subgraph "Iterations"
direction LR
C["Iteration i
initial_output"] --> D["User human_feedback"]
D --> E["improved_output"]
E --> F["Append to training_data.pkl
by agent_id and iteration"]
end
B --> C
F --> G{"More iterations?"}
G -- "Yes" --> C
G -- "No" --> H["Evaluate per agent
aggregate iterations"]
H --> I["Consolidate
suggestions[] + quality + final_summary"]
I --> J["Save by agent role to trained file
(default: trained_agents_data.pkl)"]
J --> K["Normal (non-training) runs"]
K --> L["Auto-load suggestions
from trained_agents_data.pkl"]
L --> M["Append to prompt
for consistent improvements"]
initial_output: the agent’s first answerhuman_feedback: your inline feedback when promptedimproved_output: the agent’s follow-up answer after feedbacktraining_data.pkl keyed by the agent’s internal ID and iteration.human_input = true, so running in a non-interactive environment will block on user input.train(...) finishes, CrewAI evaluates the collected training data per agent and produces a consolidated result containing:
suggestions: clear, actionable instructions distilled from your feedback and the difference between initial/improved outputsquality: a 0–10 score capturing improvementfinal_summary: a step-by-step set of action items for future taskstrain(...) (default via CLI is trained_agents_data.pkl). Entries are keyed by the agent’s role so they can be applied across sessions.suggestions and appends them to the task prompt as mandatory instructions. This gives you consistent improvements without changing your agent definitions.training_data.pkl (ephemeral, per-session):
agent_id -> { iteration_number: { initial_output, human_feedback, improved_output } }trained_agents_data.pkl (or your custom filename):
agent_role -> { suggestions: string[], quality: number, final_summary: string }-f to set a custom (including absolute) path```python
from crewai import Agent, Crew, Task, LLM
# Recommended minimum for training evaluation
llm = LLM(model="mistral/open-mistral-7b")
# Better options for reliable training evaluation
llm = LLM(model="anthropic/claude-3-sonnet-20240229-v1:0")
llm = LLM(model="gpt-4o")
# Use this LLM with your agents
agent = Agent(
role="Training Evaluator",
goal="Provide accurate training feedback",
llm=llm
)
```
<Tip>
More powerful models provide higher quality feedback with better reasoning, leading to more effective training iterations.
</Tip>
```python
# Using a smaller model (expect some limitations)
llm = LLM(model="huggingface/microsoft/Phi-3-mini-4k-instruct")
```
<Warning>
While CrewAI includes optimizations for small models, expect less reliable and less nuanced evaluation results that may require more human intervention during training.
</Warning>
n_iterations) is a positive integer. The code will raise a ValueError if this condition is not met..pkl. The code will raise a ValueError if this condition is not met.trained_agents_data.pkl located in the current working directory. If you trained to a different filename, either rename it to trained_agents_data.pkl before running, or adjust the loader in code.crewai train with -f/--filename. Absolute paths are supported if you want to save outside the CWD.It is important to note that the training process may take some time, depending on the complexity of your agents and will also require your feedback on each iteration.
Once the training is complete, your agents will be equipped with enhanced capabilities and knowledge, ready to tackle complex tasks and provide more consistent and valuable insights.
Remember to regularly update and retrain your agents to ensure they stay up-to-date with the latest information and advancements in the field.