Adamax: Difference between revisions

From Cornell University Computational Optimization Open Textbook - Optimization Wiki
Jump to navigation Jump to search
No edit summary
No edit summary
Line 11: Line 11:
The Adamax optimization algorithm follows these steps:
The Adamax optimization algorithm follows these steps:


**Step 1: Initialize Parameters**  
*Step 1: Initialize Parameters   
Set the learning rate <math>\alpha</math>, exponential decay rates <math>\beta_1</math> and <math>\beta_2</math>, and numerical stability constant <math>\epsilon</math>. Initialize the first moment estimate <math>m = 0</math> and infinity norm estimate <math>u = 0</math>.   
Set the learning rate <math>\alpha</math>, exponential decay rates <math>\beta_1</math> and <math>\beta_2</math>, and numerical stability constant <math>\epsilon</math>. Initialize the first moment estimate <math>m = 0</math> and infinity norm estimate <math>u = 0</math>.   


**Step 2: Gradient Computation**  
*Step 2: Gradient Computation   
Compute the gradient of the loss function with respect to the model parameters, <math>g_t</math>.   
Compute the gradient of the loss function with respect to the model parameters, <math>g_t</math>.   


**Step 3: Update First Moment Estimate** 
*Step 3: Update First Moment Estimate  
<math>m_t = \beta_1 m_{t-1} + (1 - \beta_1) g_t</math>   
<math>m_t = \beta_1 m_{t-1} + (1 - \beta_1) g_t</math>   


**Step 4: Update Infinity Norm Estimate**  
*Step 4: Update Infinity Norm Estimate   
<math>u_t = \max(\beta_2 u_{t-1}, |g_t|)</math>   
<math>u_t = \max(\beta_2 u_{t-1}, |g_t|)</math>   


**Step 5: Bias-Corrected Learning Rate**  
*Step 5: Bias-Corrected Learning Rate   
<math>\hat{\alpha} = \frac{\alpha}{1 - \beta_1^t}</math>   
<math>\hat{\alpha} = \frac{\alpha}{1 - \beta_1^t}</math>   


**Step 6: Parameter Update**  
*Step 6: Parameter Update   
<math>\theta_t = \theta_{t-1} - \frac{\hat{\alpha} \cdot m_t}{u_t + \epsilon}</math>   
<math>\theta_t = \theta_{t-1} - \frac{\hat{\alpha} \cdot m_t}{u_t + \epsilon}</math>   


Line 51: Line 51:
=== Step-by-Step Calculations ===
=== Step-by-Step Calculations ===


==== Iteration 1
==== Iteration 1 <math>t = 1</math> ====
<math>t = 1</math> ====
*Gradient Calculation   
*Gradient Calculation   
<math>g_1 = 2x_0 = 2 \cdot 2.0 = 4.0</math>   
<math>g_1 = 2x_0 = 2 \cdot 2.0 = 4.0</math>   
Line 78: Line 77:
The parameter moves closer to the function's minimum at <math>x = 0</math>.   
The parameter moves closer to the function's minimum at <math>x = 0</math>.   


==== Iteration <math>t = 2</math> ====
==== Iteration 2 <math>t = 2</math> ====
*Time Step Update   
*Time Step Update   
<math>t = 2</math>   
<math>t = 2</math>   

Revision as of 02:20, 15 December 2024

Author: Chengcong Xu (cx253), Jessica Liu (hl2482), Xiaolin Bu (xb58), Qiaoyue Ye (qy252), Haoru Feng (hf352) (ChemE 6800 Fall 2024)

Stewards: Nathan Preuss, Wei-Han Chen, Tianqi Xiao, Guoqing Hu

Introduction

Adamax is an optimization algorithm introduced by Kingma and Ba in their Adam optimizer paper (2014). It improves upon the Adam algorithm by replacing the second moment's root mean square (RMS) norm with the infinity norm (). This change makes Adamax more robust and numerically stable, especially when handling sparse gradients, noisy updates, or optimization problems with significant gradient variations.

Adamax dynamically adjusts learning rates for individual parameters, making it well-suited for training deep neural networks, large-scale machine learning models, and tasks involving high-dimensional parameter spaces.

Algorithm Discussion

The Adamax optimization algorithm follows these steps:

  • Step 1: Initialize Parameters

Set the learning rate , exponential decay rates and , and numerical stability constant . Initialize the first moment estimate and infinity norm estimate .

  • Step 2: Gradient Computation

Compute the gradient of the loss function with respect to the model parameters, .

  • Step 3: Update First Moment Estimate

  • Step 4: Update Infinity Norm Estimate

  • Step 5: Bias-Corrected Learning Rate

  • Step 6: Parameter Update


Numerical Examples

To illustrate the Adamax optimization algorithm, we will minimize the quadratic function with step-by-step calculations.

Problem Setup

  • Optimization Objective: Minimize , which reaches its minimum at with .
  • Initial Parameter: Start with .
  • Gradient Formula: , which determines the direction and rate of parameter change.
  • Hyperparameters:

- Learning Rate: controls the step size. - First Moment Decay Rate: , determines how past gradients influence the current gradient estimate. - Infinity Norm Decay Rate: , governs the decay of the infinity norm used for scaling updates. - Numerical Stability Constant: , prevents division by zero.

  • Initialization:

Step-by-Step Calculations

Iteration 1

  • Gradient Calculation

The gradient indicates the steepest direction and magnitude for reducing . A positive gradient shows must decrease to minimize the function.

  • First Moment Update

The first moment is a running average of past gradients, smoothing out fluctuations.

  • Infinity Norm Update

The infinity norm ensures updates are scaled by the largest observed gradient, stabilizing step sizes.

  • Bias-Corrected Learning Rate

The learning rate is corrected for bias introduced by initialization, ensuring effective parameter updates.

  • Parameter Update

The parameter moves closer to the function's minimum at .

Iteration 2

  • Time Step Update

  • Gradient Calculation

  • First Moment Update

  • Infinity Norm Update

  • Bias-Corrected Learning Rate

  • Parameter Update

The parameter continues to approach the minimum at .

Summary

Through these two iterations, Adamax effectively adjusts the parameter based on the computed gradients, moving it closer to the minimum. The use of the infinity norm stabilizes the updates, ensuring smooth convergence.


Applications

Natural Language Processing

Adamax is particularly effective in training transformer-based models like BERT and GPT. Its stability with sparse gradients makes it ideal for tasks such as text classification, machine translation, and named entity recognition.

Computer Vision

In computer vision, Adamax optimizes deep CNNs for tasks like image classification and object detection. Its smooth convergence behavior has been observed to enhance performance in models like ResNet and DenseNet.

Reinforcement Learning

Adamax has been applied in training reinforcement learning agents, particularly in environments where gradient updates are inconsistent or noisy, such as robotic control and policy optimization.

Generative Models

For training generative models, including GANs and VAEs, Adamax provides robust optimization, improving stability and output quality during adversarial training.

Time-Series Forecasting

Adamax is used in financial and economic forecasting, where it handles noisy gradients effectively, resulting in stable and accurate time-series predictions.

Advantages over Other Approaches

  • Stability: The use of the infinity norm ensures Adamax handles gradient variations smoothly.
  • Sparse Gradient Handling: Adamax is robust in scenarios with zero or near-zero gradients, common in NLP tasks.
  • Efficiency: Adamax is computationally efficient for high-dimensional optimization problems.

Conclusion

Adamax is a robust and efficient variant of the Adam optimizer that replaces the RMS norm with the infinity norm. Its ability to handle sparse gradients, noisy updates, and large parameter spaces makes it a widely used optimization method in natural language processing, computer vision, reinforcement learning, and generative modeling.

Future advancements may involve integrating Adamax with learning rate schedules and regularization techniques to further enhance its performance.

References