Citation Relationships



Amari S (2007) Integration of stochastic models by minimizing alpha-divergence. Neural Comput 19:2780-96 [PubMed]

References and models cited by this paper

References and models that cite this paper

Amari S, Ikeda S, Shimokawa H (2001) Information geometry and meanfield approximation: The Œ+ projection approach Advanced mean field methods: Theory and practive, Opper M:Saad D, ed. pp.241

Amari S, Nagaoka H (2000) Methods of information geometry

Chernoff H (1952) A measure of asymptotic efficiency for tests of a hypothesis based on a sum of observations Ann Math Stat 23:493-507

Corcuera JM, Giummole F (1999) A generalized Bayes rule for prediction Scandinavian J Stat 26:265-279

Csiszar I (1975) I-divergence geometry of probability distributions and minimization problems Annals Of Probability 3:146-158

Dayan P, Hinton GE, Neal RM, Zemel RS (1995) The Helmholtz machine. Neural Comput 7:889-904 [PubMed]

Eguchi S (1983) Second order efficiency of minimum contrast estimators in a curved exponential family Ann Stat 11:793-803

Falmagne JC (1985) Elements of psychophysical theory

Hardy G, Littlewood JE, Polya G (1952) Inequalities

Heller J (2006) Illumination-invariance of Plateaus midgray J Math Psychol 50:263-270

Hida E, Ohno H, Hashimoto N, Amari S, Saito H (2006) Neural representation for perception of wide-field visual flow in MST: Bidirectional transparent motion and its illusory after effect Manuscript submitted for publication

Hinton GE (2002) Training products of experts by minimizing contrastive divergence. Neural Comput 14:1771-800 [Journal] [PubMed]

Ikeda S, Tanaka T, Amari S (2004) Stochastic reasoning, free energy, and information geometry. Neural Comput 16:1779-810 [Journal] [PubMed]

Jacobs RA, Jordan MI, Nowlan SJ, Hinton GE (1991) Adaptive mixtures of local experts Neural Comput 3:79-87

Jordan MI, Jacobs RA (1994) Hierarchical mixtures of experts and the EM algorithm Neural Comput 6:181-214

Komaki F (1996) On asymptotic properties of predictive distributions Biometrika 83:299-313

Marriott P (2002) On the local geometry of mixture models Biometrika 89:79-93

Matsuyama Y (2003) The alpha-EM algorithm: Surrogate likelihood maximization using alpha-logarithmic information measures. IEEE Trans Inform Theory 49:692-706

Mihoko M, Eguchi S (2002) Robust blind source separation by beta divergence. Neural Comput 14:1859-86 [Journal] [PubMed]

Minka T (2005) Divergence measures and message passing MSR-TR-2005-173

Murata N, Takenouchi T, Kanamori T, Eguchi S (2004) Information geometry of U-Boost and Bregman divergence. Neural Comput 16:1437-81 [Journal] [PubMed]

Petz D, Temesi R (2005) Means of positive numbers and matrices SIAM J Matrix Analysis and Applications 27:712-720

Renyi H (1961) On measures of entropy and information Proc 4th Berkeley Symp Math Stat Prob

Toyoizumi T, Aihara K (2006) Generalization of the mean-field method for power-law distributions Intl J Bifurcation Chaos 16:129-135

Tsallis C (1988) Possible generalization of Boltzmann-Gibbs statistics J Stat Phys 52:479-487

Wolpert DM, Kawato M (1998) Multiple paired forward and inverse models for motor control. Neural Netw 11:1317-29 [PubMed]

Xu L (2004) Advances on BYY harmony learning: information theoretic perspective, generalized projection geometry, and independent factor autodetermination. IEEE Trans Neural Netw 15:885-902 [Journal] [PubMed]

Zhang J (2004) Divergence function, duality, and convex analysis. Neural Comput 16:159-95 [PubMed]

Zhu H, Rohwer R (1998) Information geometry, Bayesian inference, ideal estimates, and error decomposition Unpublished manuscript. Available online at http:--www.santafe.edu-research-publications-wpabstract-199806045

(29 refs)