Citation Relationships

Legends: Link to a Model Reference cited by multiple papers

Amari S (2007) Integration of stochastic models by minimizing alpha-divergence. Neural Comput 19:2780-96 [PubMed]

References and models cited by this paper

References and models that cite this paper

Amari S, Ikeda S, Shimokawa H (2001) Information geometry and meanfield approximation: The Œ+ projection approach Advanced mean field methods: Theory and practive, Opper M:Saad D, ed. pp.241
Amari S, Nagaoka H (2000) Methods of information geometry
Chernoff H (1952) A measure of asymptotic efficiency for tests of a hypothesis based on a sum of observations Ann Math Stat 23:493-507
Corcuera JM, Giummole F (1999) A generalized Bayes rule for prediction Scandinavian J Stat 26:265-279
Csiszar I (1975) I-divergence geometry of probability distributions and minimization problems Annals Of Probability 3:146-158
Dayan P, Hinton GE, Neal RM, Zemel RS (1995) The Helmholtz machine. Neural Comput 7:889-904 [PubMed]
Eguchi S (1983) Second order efficiency of minimum contrast estimators in a curved exponential family Ann Stat 11:793-803
Falmagne JC (1985) Elements of psychophysical theory
Heller J (2006) Illumination-invariance of Plateaus midgray J Math Psychol 50:263-270
Hida E, Ohno H, Hashimoto N, Amari S, Saito H (2006) Neural representation for perception of wide-field visual flow in MST: Bidirectional transparent motion and its illusory after effect Manuscript submitted for publication
Hinton GE (2002) Training products of experts by minimizing contrastive divergence. Neural Comput 14:1771-800 [Journal] [PubMed]
Ikeda S, Tanaka T, Amari S (2004) Stochastic reasoning, free energy, and information geometry. Neural Comput 16:1779-810 [Journal] [PubMed]
Jacobs RA, Jordan MI, Nowlan SJ, Hinton GE (1991) Adaptive mixtures of local experts Neural Comput 3:79-87
Jordan MI, Jacobs RA (1994) Hierarchical mixtures of experts and the EM algorithm Neural Comput 6:181-214
Komaki F (1996) On asymptotic properties of predictive distributions Biometrika 83:299-313
Marriott P (2002) On the local geometry of mixture models Biometrika 89:79-93
Matsuyama Y (2003) The alpha-EM algorithm: Surrogate likelihood maximization using alpha-logarithmic information measures. IEEE Trans Inform Theory 49:692-706
Mihoko M, Eguchi S (2002) Robust blind source separation by beta divergence. Neural Comput 14:1859-86 [Journal] [PubMed]
Minka T (2005) Divergence measures and message passing MSR-TR-2005-173
Murata N, Takenouchi T, Kanamori T, Eguchi S (2004) Information geometry of U-Boost and Bregman divergence. Neural Comput 16:1437-81 [Journal] [PubMed]
Petz D, Temesi R (2005) Means of positive numbers and matrices SIAM J Matrix Analysis and Applications 27:712-720
Renyi H (1961) On measures of entropy and information Proc 4th Berkeley Symp Math Stat Prob
Toyoizumi T, Aihara K (2006) Generalization of the mean-field method for power-law distributions Intl J Bifurcation Chaos 16:129-135
Tsallis C (1988) Possible generalization of Boltzmann-Gibbs statistics J Stat Phys 52:479-487
Wolpert DM, Kawato M (1998) Multiple paired forward and inverse models for motor control. Neural Netw 11:1317-29 [PubMed]
Xu L (2004) Advances on BYY harmony learning: information theoretic perspective, generalized projection geometry, and independent factor autodetermination. IEEE Trans Neural Netw 15:885-902 [Journal] [PubMed]
Zhang J (2004) Divergence function, duality, and convex analysis. Neural Comput 16:159-95 [PubMed]
Zhu H, Rohwer R (1998) Information geometry, Bayesian inference, ideal estimates, and error decomposition Unpublished manuscript. Available online at
(29 refs)