You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I observed that the hyperparameter offset (introduced as downweighting factor in aa56561) which corresponds to tau_0 from Hoffman et al. is set differently compared to the original algorithm proposed by Hoffman.
self._tau0 = tau0 + 1
When passing tau0 to his algorithm, tau0 + 1 is actually used in the calculation of rhot:
I find this resource helpful: https://vb-learning-rate-demo.herokuapp.com to see how the parameters offset (tau) and decay (kappa) affect the learning rate of the online VB method
I observed that the hyperparameter
offset
(introduced asdownweighting
factor in aa56561) which corresponds to tau_0 from Hoffman et al. is set differently compared to the original algorithm proposed by Hoffman.When passing
tau0
to his algorithm,tau0 + 1
is actually used in the calculation of rhot:This line is used here and here.
The commit edc3ce5 in gensim further changes the computation of
rho
(in order to pay attention to multi-pass algorithm as discussed in #298:https://github.com/RaRe-Technologies/gensim/blob/351456b4f7d597e5a4522e71acedf785b2128ca1/gensim/models/ldamodel.py#L963-L967
I wonder if there is any rationale behind this decision to deviate from Hoffman's tau0 or if this was unintended?
The text was updated successfully, but these errors were encountered: