Created
March 4, 2015 09:12
-
-
Save skaae/ae7225263ca8806868cb to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
def adam(loss, all_params, learning_rate=0.001, b1=0.9, b2=0.999, e=1e-8, | |
gamma=1-1e-8): | |
""" | |
ADAM update rules | |
Default values are taken from [Kingma2014] | |
References: | |
[Kingma2014] Kingma, Diederik, and Jimmy Ba. | |
"Adam: A Method for Stochastic Optimization." | |
arXiv preprint arXiv:1412.6980 (2014). | |
http://arxiv.org/pdf/1412.6980v4.pdf | |
""" | |
updates = [] | |
all_grads = theano.grad(loss, all_params) | |
alpha = learning_rate | |
t = theano.shared(np.float32(1)) | |
b1_t = b1*gamma**(t-1) #(Decay the first moment running average coefficient) | |
for theta_previous, g in zip(all_params, all_grads): | |
m_previous = theano.shared(np.zeros(theta_previous.get_value().shape, | |
dtype=theano.config.floatX)) | |
v_previous = theano.shared(np.zeros(theta_previous.get_value().shape, | |
dtype=theano.config.floatX)) | |
m = b1_t*m_previous + (1 - b1_t)*g # (Update biased first moment estimate) | |
v = b2*v_previous + (1 - b2)*g**2 # (Update biased second raw moment estimate) | |
m_hat = m / (1-b1**t) # (Compute bias-corrected first moment estimate) | |
v_hat = v / (1-b2**t) # (Compute bias-corrected second raw moment estimate) | |
theta = theta_previous - (alpha * m_hat) / (T.sqrt(v_hat) + e) #(Update parameters) | |
updates.append((m_previous, m)) | |
updates.append((v_previous, v)) | |
updates.append((theta_previous, theta) ) | |
updates.append((t, t + 1.)) | |
return updates |
Why is m_previous always resetted to 0? I thought 'm' (and 'v') should retain the value from previous iterations?
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I get good results with this Adam implementation, and I do not get good results with other impls (e.g. Alec Radford). The only (functional) change I see is line 18 (gamma). What/who is the source I can cite for this? e.g. I do not see this in Algorithm 1 in the original paper.