From Parameter Learning to Prompt Learning: A Parallel Evolution

Nov. 2025

Key Insight: The evolution of prompt optimization mirrors the historical development of neural network training—from simple perturbation methods to sophisticated gradient-based techniques. Understanding this parallel can help us anticipate future directions in prompt engineering.

1. Introduction

Large Language Models (LLMs) have revolutionized how we approach natural language tasks. However, their performance remains highly sensitive to one critical factor: the quality of the prompts we provide them. Just as neural networks once required careful manual tuning of architectures and hyperparameters, LLMs today demand careful prompt engineering—a process that can be time-consuming, expertise-dependent, and frustratingly trial-and-error.

Interestingly, the field of prompt optimization is now retracing the same evolutionary path that parameter learning followed decades ago. This parallel is not merely coincidental—it reflects fundamental principles about how we optimize in discrete versus continuous spaces, and how we balance exploration with exploitation.

Figure 1: the Comparison between the Evolution of Parameter Optimization and Prompt Optimization

Left: Parameter Learning Evolution (1980s Genetic Algorithms → 1990s SGD → 2000s Adam/Advanced optimizers) Right: Prompt Learning Evolution (2022 Genetic approaches → 2023 Textual gradients → 2024 Advanced methods)

2. The Parameter Learning Story: Setting the Stage

Before diving into prompt optimization, let's briefly revisit the evolution of parameter learning in neural networks—a journey that spanned several decades and fundamentally transformed machine learning.

2.1 The Era of Genetic Algorithms

In the early days of neural networks, researchers faced a challenging problem: how to find good parameter values in high-dimensional spaces without gradient information (or with gradients that were difficult to compute reliably). The solution many turned to was inspired by biological evolution: genetic algorithms.

The core idea was elegantly simple:

  1. Maintain a population of candidate solutions (parameter sets)
  2. Evaluate each candidate's fitness (performance on the task)
  3. Select the best-performing candidates
  4. Create new candidates through mutation and crossover
  5. Repeat until convergence

These methods were gradient-free—they treated the model as a black box and relied purely on performance feedback. While they could escape local optima and didn't require differentiability, they were computationally expensive, requiring many function evaluations.

2.2 The Gradient Descent Revolution

The landscape changed dramatically with the realization that we could efficiently compute gradients through backpropagation. Instead of exploring randomly, we could follow the direction of steepest descent. The standard parameter update became:

$$\theta_{t+1} = \theta_t - \eta \nabla_\theta L(\theta_t)$$

where \(\theta\) represents the parameters, \(\eta\) is the learning rate, and \(\nabla_\theta L\) is the gradient of the loss with respect to parameters. This first-order method was transformative—it was efficient, directed, and mathematically principled.

2.3 Beyond First-Order: Advanced Optimizers

But first-order gradient descent had limitations: sensitivity to learning rates, slow convergence in ill-conditioned spaces, and susceptibility to saddle points. This led to a new generation of optimizers:

Momentum-Based Methods

These methods maintain a moving average of past gradients:

$$v_{t+1} = \beta v_t + \nabla_\theta L(\theta_t)$$ $$\theta_{t+1} = \theta_t - \eta v_{t+1}$$

This helps accelerate convergence and dampen oscillations.

Adam and Adaptive Methods

Adam combines momentum with adaptive learning rates:

$$m_t = \beta_1 m_{t-1} + (1-\beta_1) \nabla_\theta L$$ $$v_t = \beta_2 v_{t-1} + (1-\beta_2) (\nabla_\theta L)^2$$ $$\theta_{t+1} = \theta_t - \eta \frac{m_t}{\sqrt{v_t} + \epsilon}$$

This provides per-parameter adaptive learning rates while maintaining momentum.

Second-Order Methods

Methods like Newton's method use curvature information:

$$\theta_{t+1} = \theta_t - H^{-1} \nabla_\theta L(\theta_t)$$

where \(H\) is the Hessian matrix. While computationally expensive, these methods can converge much faster in certain settings.

Figure 2: GPS algorithm from (Xu et al., 2022)

3. The Prompt Learning Parallel: History Repeating

Now, let's see how the prompt optimization community is essentially walking the same path—but in the discrete space of natural language.

3.1 Why Prompt Optimization is Challenging

Prompt optimization faces a unique challenge: prompts are discrete text sequences that must remain coherent and human-readable. We can't simply compute \(\nabla_p L(p)\) where \(p\) is a prompt—gradients with respect to discrete text don't have a natural definition. This forces us to work in a fundamentally different optimization landscape.

The Key Difference: In parameter learning, we optimize in continuous \(\mathbb{R}^n\) space. In prompt learning, we optimize in a discrete space of coherent language expressions. This changes everything about how we can approach optimization.

3.2 Phase 1: Genetic Approaches to Prompt Search

Given the discrete nature of prompts, it's perhaps unsurprising that the field first turned to evolutionary methods—exactly as parameter learning did in its early days.

3.2.1 GPS: Genetic Prompt Search

GPS (Xu et al., 2022) directly applies genetic algorithm principles to prompt optimization:

  • Population: A set of candidate prompts \(P = \{p_1, p_2, ..., p_N\}\)
  • Fitness: Performance on a small validation set
  • Selection: Keep top-K performing prompts
  • Mutation: Use LLMs or back-translation to create variations
  • Crossover: Combine parts of different successful prompts

The algorithm is gradient-free and requires no parameter updates—just like early genetic algorithms for neural networks. GPS demonstrates that simple perturbation and selection can improve prompts by 2.6 points over manual baselines.

Figure 3: the SoS framework from (Sinha et al., 2024)

SoS framework diagram showing: Multi-objective evolution strategy, Semantic, feedback, and crossover mutations, Security vs Performance tradeoff visualization. This shows evolution of genetic methods to handle multiple objectives.

3.2.2 SoS: Multi-Objective Evolution

Survival of the Safest (Sinha et al., 2024) extends the genetic approach to handle multiple objectives simultaneously—particularly balancing performance with security:

$$p^* = \arg\max_{p \in \mathcal{P}} \left[ w_1 \cdot \text{Performance}(p) + w_2 \cdot \text{Security}(p) \right]$$

SoS introduces an interleaved evolution strategy that:

  • Alternates between optimizing different objectives
  • Uses semantic mutations to maintain prompt coherence
  • Provides users with a pool of Pareto-optimal candidates

This approach builds on genetic foundations while adding sophistication for real-world deployment concerns—a natural evolution of the paradigm.

3.2.3 EvoPrompt: Connecting LLMs with Evolution

EvoPrompt (Guo et al., 2024) makes a key insight: we can use LLMs themselves as intelligent mutation operators. Rather than random perturbations, the system:

  • Uses LLMs to generate semantically meaningful variations
  • Implements evolutionary operators (mutation, crossover) through natural language instructions
  • Maintains population diversity while ensuring prompt quality

This represents a maturation of genetic approaches—the mutations are no longer random but guided by the linguistic intelligence of LLMs. Yet the core framework remains evolutionary: maintain a population, evaluate fitness, select, and reproduce.

Method Selection Strategy Mutation Approach Key Innovation
GPS Top-K by validation performance Back-translation, random edits First gradient-free prompt search
SoS Multi-objective Pareto selection Semantic + crossover mutations Balances performance and security
EvoPrompt Fitness-proportional selection LLM-guided intelligent mutations Uses LLMs as mutation operators

3.3 Phase 2: The Gradient Revolution in Text Space

Just as parameter learning evolved from genetic algorithms to gradient methods, prompt optimization is now undergoing its own "gradient revolution." But how do we compute gradients in discrete text space?

3.3.1 ProTeGi: Textual Gradients with Beam Search

ProTeGi (Pryzant et al., 2023) introduces the concept of "textual gradients"—natural language feedback that plays an analogous role to numerical gradients:

  1. Forward Pass: Run the current prompt on a mini-batch of data
  2. Compute "Gradient": Use an LLM to analyze errors and generate natural language criticism:
    $$\nabla_p L \approx \text{LLM}(\text{"What is wrong with this prompt given these errors?"})$$
  3. Backpropagation: Edit the prompt in the direction suggested by the criticism
  4. Beam Search: Maintain multiple candidate prompts and select the best

This is analogous to first-order gradient descent, but operating through natural language feedback rather than numerical derivatives. The method demonstrates improvements of up to 31% over initial prompts.

3.3.2 TextGrad: Automatic Differentiation via Text

TextGrad (Yuksekgonul et al., 2024) takes the gradient analogy further by building a complete automatic differentiation framework for text:

$$\frac{\partial \mathcal{L}}{\partial p} = \text{TextGrad}\left(\text{prompt } p, \text{ error context}\right)$$

The key innovations include:

  • Computation Graphs: Model compound AI systems as directed acyclic graphs
  • Backpropagation Through Modules: Propagate textual feedback backwards through the system
  • PyTorch-like API: Familiar syntax for researchers used to modern deep learning frameworks

TextGrad enables optimization of not just prompts but entire AI systems composed of multiple LLM calls, tools, and data processing steps. It treats each component as differentiable with respect to textual feedback.

-->

Figure 4: the concept of TextGrad

The Gradient Analogy: While these aren't true mathematical gradients, textual feedback serves the same purpose: it provides direction on how to modify the input (prompt) to reduce the error (improve performance). The key difference is that the "direction" is expressed in natural language rather than a numerical vector.

3.4 Phase 3: Beyond First-Order—Advanced Prompt Optimization

Just as parameter learning didn't stop at vanilla gradient descent, prompt optimization is now developing more sophisticated methods that go beyond simple first-order textual gradients.

3.4.1 REVOLVE: Tracking Response Evolution

REVOLVE (Zhang et al., 2024) represents a significant advance by tracking how responses evolve across iterations—analogous to momentum and second-order methods in numerical optimization.

The key insight is that first-order methods (like vanilla TextGrad) only use immediate feedback from the current iteration. REVOLVE extends this by considering:

$$\text{Gradient}_{\text{REVOLVE}} = \text{ImmediateFeedback}(p_t) + \text{EvolutionFeedback}(p_t, p_{t-1}, ...)$$

Specifically, REVOLVE:

  • Tracks Response Patterns: Monitors how model outputs change across iterations
  • Detects Stagnation: Identifies when improvements have plateaued
  • Adaptive Adjustments: Makes larger changes when stuck in local optima, smaller changes when converging

This is reminiscent of how momentum methods in SGD use historical gradient information to accelerate convergence and escape saddle points:

$$v_t = \beta v_{t-1} + \nabla L_t \quad \text{(momentum in parameter space)}$$ $$\Delta p_t = f(\text{history of responses}) \quad \text{(REVOLVE in prompt space)}$$

Results show REVOLVE achieves 7.8% improvement in prompt optimization, 20.72% in solution refinement, and converges in fewer iterations than first-order methods.

-->

Figure 5: the concept of TextGrad extended to the break local optimal in optimization.

3.4.2 SIPDO: Closed-Loop with Synthetic Data

SIPDO (Yu et al., 2025) introduces another sophisticated mechanism: generating synthetic data to actively probe prompt weaknesses. This creates a closed feedback loop:

  1. Synthetic Data Generation: Create examples specifically designed to challenge the current prompt
  2. Prompt Optimization: Refine the prompt based on failures on synthetic data
  3. Difficulty Progression: Gradually increase synthetic data difficulty
  4. Iterate: The improved prompt influences what synthetic data is generated next

The objective function becomes:

$$\min_{\psi} \mathcal{R}(\psi) + \lambda \mathbb{E}_{(\tilde{x},\tilde{y}) \sim q_\psi} \left[ L(f(p, \tilde{x}), \tilde{y}) \right]$$

where \(q_\psi\) is the synthetic data distribution and \(\mathcal{R}(\psi)\) regularizes it to match the true label distribution.

This approach is analogous to adversarial training or curriculum learning in deep learning—the system actively seeks out its weaknesses rather than passively responding to a fixed dataset.

-->

Figure 6: SIPDO uses generated synthetic data to optimize prompts.

4. Comparative Analysis: The Parallel Patterns

Let's now step back and explicitly compare the evolution in both domains:

Era Parameter Learning Prompt Learning Common Principles
Phase 1
Exploration-Based
• Genetic Algorithms
• Random Search
• Simulated Annealing
• GPS
• SoS
• EvoPrompt
• Gradient-free
• Population-based
• Exploration emphasis
• No differentiability required
Phase 2
Gradient-Based
• SGD
• Backpropagation
• First-order optimization
• ProTeGi
• TextGrad
• Textual gradients
• Directed optimization
• Efficient convergence
• Local feedback
• Exploitation emphasis
Phase 3
Advanced Methods
• Momentum
• Adam
• Second-order methods
• Adaptive learning rates
• REVOLVE
• SIPDO
• Response evolution tracking
• Synthetic data feedback
• Historical information
• Adaptive step sizes
• Escape local optima
• Accelerated convergence

4.1 Why the Parallel?

This parallel evolution isn't coincidental—it reflects fundamental principles of optimization:

  1. Start Simple: When facing a new optimization problem, we naturally begin with methods that make few assumptions (like evolutionary approaches).
  2. Exploit Structure: As we understand the problem better, we can leverage its structure (like using feedback that approximates gradients).
  3. Refine and Accelerate: Finally, we develop sophisticated methods that use richer information (historical patterns, curvature, adaptive strategies).

The key difference is time scale: parameter learning took decades to evolve from genetic algorithms to Adam; prompt learning is compressing this journey into just a few years, benefiting from our accumulated understanding of optimization principles.

5. Future Perspectives: What's Next for Prompt Optimization?

If the parallel continues, what might we expect next in prompt optimization?

5.1 Potential Directions

1. True Second-Order Methods for Prompts

Could we develop methods that explicitly model the "curvature" of the prompt space? This might involve:

  • Estimating how sensitive performance is to different types of prompt changes
  • Building surrogate models that predict prompt performance
  • Using meta-learning to learn how prompts should be updated

2. Adaptive "Learning Rates" for Prompt Updates

Just as Adam adapts learning rates per parameter, we might develop methods that:

  • Learn which parts of a prompt are most sensitive to changes
  • Adapt the magnitude of edits based on optimization progress
  • Use different update strategies for different prompt components

3. Prompt Optimization Preconditioning

Analogous to preconditioned gradient methods, we might:

  • Learn transformations of the prompt space that make optimization easier
  • Develop better initializations based on task characteristics
  • Transfer optimization strategies across related tasks

4. Multi-Task Prompt Optimization

Extensions of SIPDO and REVOLVE might:

  • Optimize prompts jointly across related tasks
  • Learn meta-prompts that can be quickly adapted
  • Develop prompt representations that transfer across domains

5.2 Open Questions

Several fascinating questions remain:

  • Theoretical Foundations: Can we develop convergence guarantees for textual gradient methods? Under what conditions do they converge to optimal prompts?
  • Prompt Space Geometry: What is the structure of the prompt space? Are there natural metrics or manifolds that could guide optimization?
  • Generalization: How do prompts optimized on one dataset generalize to others? Can we develop regularization techniques for prompts?
  • Efficiency: Current methods require many LLM calls. Can we develop more sample-efficient approaches, perhaps using smaller models to guide optimization for larger ones?
  • Multi-Modal Prompts: As models become multi-modal, how do we optimize prompts that include images, code, and text together?

6. Conclusion

The evolution of prompt optimization beautifully mirrors the historical development of parameter learning in neural networks. From genetic algorithms to gradient methods to advanced optimizers, we're watching history repeat itself—but this time in the discrete space of natural language.

This parallel offers both practical guidance and theoretical insight. Practically, it suggests that techniques developed for parameter optimization (momentum, adaptive learning rates, second-order methods) may have valuable analogs in prompt space. Theoretically, it highlights that good optimization principles transcend the specific domain—whether we're optimizing continuous parameters or discrete text.

The rapid pace of progress in prompt optimization, compressed into just a few years what took decades in parameter learning, demonstrates both the maturity of our optimization understanding and the unique challenges of discrete language spaces. Methods like REVOLVE and SIPDO are showing us that prompt optimization isn't just about copying techniques from parameter learning—it's about adapting fundamental principles to a new domain while developing innovations unique to the structure of language.

As LLMs continue to improve and become more central to AI systems, prompt optimization will only grow in importance. By understanding this parallel evolution, we can better anticipate future developments and contribute to building more effective, efficient, and principled methods for optimizing these powerful but sensitive systems.

Takeaway: The next breakthrough in prompt optimization might come from asking: "What would the Adam optimizer look like for prompts?" or "How would we implement natural gradient descent in text space?" The parallel isn't just historical curiosity—it's a roadmap for innovation.

References

• Xu, H., Chen, Y., Du, Y., et al. (2022). GPS: Genetic Prompt Search for Efficient Few-shot Learning.
• Sinha, A., Cui, W., Das, K., & Zhang, J. (2024). Survival of the Safest: Towards Secure Prompt Optimization through Interleaved Multi-Objective Evolution.
• Guo, Q., Wang, R., Guo, J., et al. (2024). EvoPrompt: Connecting LLMs with Evolutionary Algorithms Yields Powerful Prompt Optimizers.
• Pryzant, R., Iter, D., Li, J., et al. (2023). Automatic Prompt Optimization with "Gradient Descent" and Beam Search.
• Yuksekgonul, M., Bianchi, F., Boen, J., et al. (2024). TextGrad: Automatic "Differentiation" via Text.
• Zhang, P., Jin, H., Hu, L., et al. (2024). REVOLVE: Optimizing AI Systems by Tracking Response Evolution in Textual Optimization.
• Yu, Y., Yu, Y., Wei, K., et al. (2025). SIPDO: Closed-Loop Prompt Optimization via Synthetic Data Feedback.