This shows you the differences between two versions of the page.
| Next revision | Previous revision | ||
| fireact_agent_finetuning [2026/03/25 15:22] – Create FireAct page: multi-task agent fine-tuning enabling small models to rival GPT-4 agent | fireact_agent_finetuning [2026/03/30 22:20] (current) – Restructure: footnotes as references agent | ||
|---|---|---|---|
| Line 1: | Line 1: | ||
| ====== FireAct: Toward Language Agent Fine-tuning ====== | ====== FireAct: Toward Language Agent Fine-tuning ====== | ||
| - | FireAct is a fine-tuning approach that enables **smaller language models to perform agentic tasks at levels approaching GPT-4** by training on diverse trajectories generated by stronger models. Introduced by Chen et al. (2023), FireAct demonstrates that multi-task, multi-method trajectory data is the key to effective agent fine-tuning. | + | FireAct is a fine-tuning approach that enables **smaller language models to perform agentic tasks at levels approaching GPT-4** by training on diverse trajectories generated by stronger models. Introduced by Chen et al. (2023), FireAct demonstrates that multi-task, multi-method trajectory data is the key to effective agent fine-tuning.(([[https:// |
| ===== Overview ===== | ===== Overview ===== | ||
| Line 95: | Line 95: | ||
| trainer.train() | trainer.train() | ||
| </ | </ | ||
| - | |||
| - | ===== References ===== | ||
| - | |||
| - | * [[https:// | ||
| - | * [[https:// | ||
| - | * [[https:// | ||
| ===== See Also ===== | ===== See Also ===== | ||
| Line 109: | Line 103: | ||
| * [[retroformer|Retroformer: | * [[retroformer|Retroformer: | ||
| + | ===== References ===== | ||