From 4006467c5fa09e06f40eab551d2c7d50a8c40f68 Mon Sep 17 00:00:00 2001 From: "A. Unique TensorFlower" Date: Fri, 1 Mar 2019 17:57:19 -0800 Subject: [PATCH] Polish the tf.optimizers.Adagrad class documentation a bit. Before: https://www.tensorflow.org/versions/r2.0/api_docs/python/tf/optimizers/Adagrad PiperOrigin-RevId: 236410591 --- .../python/keras/optimizer_v2/adagrad.py | 18 +++++++++--------- 1 file changed, 9 insertions(+), 9 deletions(-) diff --git a/tensorflow/python/keras/optimizer_v2/adagrad.py b/tensorflow/python/keras/optimizer_v2/adagrad.py index 0840aa6fae5..c30c0661506 100644 --- a/tensorflow/python/keras/optimizer_v2/adagrad.py +++ b/tensorflow/python/keras/optimizer_v2/adagrad.py @@ -41,18 +41,18 @@ class Adagrad(optimizer_v2.OptimizerV2): the smaller the updates. Initialization: + $$accum_{g_0} := \text{initial_accumulator_value}$$ - $$accum_g_0 := initial_accumulator_value$$ - + Update step: $$t := t + 1$$ - $$accum_g_t := accum_g_{t-1} + g * g$$ - $$theta_t := theta_{t-1} - lr * g / (\sqrt{accum_g_t} + \epsilon)$$ + $$accum_{g_t} := accum_{g_{t-1}} + g^2$$ + $$\theta_t := \theta_{t-1} - lr * g / (\sqrt{accum_{g_t}} + \epsilon)$$ - References - See [paper] - (http://www.jmlr.org/papers/volume12/duchi11a/duchi11a.pdf) - or this - [intro](https://ppasupat.github.io/a9online/uploads/proximal_notes.pdf). + References: + + * [Paper](http://www.jmlr.org/papers/volume12/duchi11a/duchi11a.pdf). + * [Introduction] + (https://ppasupat.github.io/a9online/uploads/proximal_notes.pdf). """ def __init__(self,