Tag

Gradient Descent

All articles tagged with #gradient descent

science-and-technology2 years ago

"Accelerating Optimization: The Power of Risky Giant Steps"

Optimization researchers have discovered that taking larger steps in gradient descent algorithms can lead to faster convergence to the optimal solution. Traditionally, the field has favored smaller steps, but recent studies have shown that larger steps can be more efficient. Computer simulations have revealed that sequences of steps with a large central leap can arrive at the optimal point nearly three times faster than constant baby steps. However, these findings are limited to smooth and convex functions, which are less relevant in practical applications. The research raises questions about the underlying structure governing the best solutions and the potential for further optimization techniques.

artificial-intelligence2 years ago

Microsoft AI Research Develops APO Framework for LLM Prompt Optimization

Microsoft AI Research has introduced Automatic Prompt Optimization (APO), a general and nonparametric prompt optimization algorithm inspired by numerical gradient descent. APO automates and improves the process of prompt development for Large Language Models (LLMs) by using mini-batches of training data to obtain natural language "gradients" that describe the flaws in a given prompt. APO consistently outperformed the baselines on various NLP tasks, including jailbreak detection, hate speech detection, fake news detection, and sarcasm detection, highlighting its potential to raise the efficiency of big language models.