2025 OptimizingGenerativeAIbyBackpro

From GM-RKB
Jump to navigation Jump to search

Subject Headings: TextGrad.

Notes

Cited By

Quotes

Abstract

Recent breakthroughs in artificial intelligence (AI) are increasingly driven by systems orchestrating multiple large language models (LLMs) and other specialized tools, such as search engines and simulators. So far, these systems are primarily handcrafted by domain experts and tweaked through heuristics rather than being automatically optimized, presenting a substantial challenge to accelerating progress. The development of artificial neural networks faced a similar challenge until backpropagation and automatic differentiation transformed the field by making optimization turnkey. Analogously, here we introduce TextGrad, a versatile framework that performs optimization by backpropagating LLM-generated feedback to improve AI systems. By leveraging natural language feedback to critique and suggest improvements to any part of a system—from prompts to outputs such as molecules or treatment plansTextGrad enables the automatic optimization of generative AI systems across diverse tasks. We demonstrate TextGrad’s generality and effectiveness through studies in solving PhD-level science problems, optimizing plans for radiotherapy treatments, designing molecules with specific properties, coding, and optimizing agentic systems. TextGrad empowers scientists and engineers to easily develop impactful generative AI systems.

References

;

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2025 OptimizingGenerativeAIbyBackproCarlos Guestrin
Mert Yuksekgonul
James Zou
Federico Bianchi
Joseph Boen
Sheng Liu
Pan Lu
Zhi Huang
Optimizing Generative AI by Backpropagating Language Model Feedback10.1038/s41586-025-08661-42025